diff options
Diffstat (limited to 'nixpkgs/pkgs/applications/networking/cluster')
3 files changed, 110 insertions, 54 deletions
diff --git a/nixpkgs/pkgs/applications/networking/cluster/cloudfoundry-cli/default.nix b/nixpkgs/pkgs/applications/networking/cluster/cloudfoundry-cli/default.nix index 0371e8c813bd..2204a9405027 100644 --- a/nixpkgs/pkgs/applications/networking/cluster/cloudfoundry-cli/default.nix +++ b/nixpkgs/pkgs/applications/networking/cluster/cloudfoundry-cli/default.nix @@ -2,15 +2,15 @@ buildGoModule rec { pname = "cloudfoundry-cli"; - version = "8.7.3"; + version = "8.7.4"; src = fetchFromGitHub { owner = "cloudfoundry"; repo = "cli"; rev = "v${version}"; - sha256 = "sha256-2ABsxoGRRUfa09tVPmn1IXDR2IXIewg/b/fmQnaKLoY="; + sha256 = "sha256-W4+2ugRSSP3HgmyQJKGCPMX7cmE7Fk3iovBOgBen+q8="; }; - vendorHash = "sha256-k2NI9zyeQM4PJo2wE3WkG5sntJGISwmz4xqQVChu8WQ="; + vendorHash = "sha256-klbKL/c7L7kHPadDa/FkpuAgHYQmuLQK6yFhph52KsU="; subPackages = [ "." ]; diff --git a/nixpkgs/pkgs/applications/networking/cluster/hadoop/containerExecutor.nix b/nixpkgs/pkgs/applications/networking/cluster/hadoop/containerExecutor.nix new file mode 100644 index 000000000000..7d5d2918e9b9 --- /dev/null +++ b/nixpkgs/pkgs/applications/networking/cluster/hadoop/containerExecutor.nix @@ -0,0 +1,37 @@ +{ version, stdenv, fetchurl, lib, cmake, openssl, platformAttrs, ... }: + +stdenv.mkDerivation (finalAttrs: { + pname = "hadoop-yarn-containerexecutor"; + inherit version; + + src = fetchurl { + url = "mirror://apache/hadoop/common/hadoop-${finalAttrs.version}/hadoop-${finalAttrs.version}-src.tar.gz"; + hash = platformAttrs.${stdenv.system}.srcHash; + }; + sourceRoot = "hadoop-${finalAttrs.version}-src/hadoop-yarn-project/hadoop-yarn/" + +"hadoop-yarn-server/hadoop-yarn-server-nodemanager/src"; + + nativeBuildInputs = [ cmake ]; + buildInputs = [ openssl ]; + cmakeFlags = [ "-DHADOOP_CONF_DIR=/run/wrappers/yarn-nodemanager/etc/hadoop" ]; + + installPhase = '' + mkdir $out + mv target/var/empty/local/bin $out/ + ''; + + meta = with lib; { + homepage = "https://hadoop.apache.org/"; + description = "Framework for distributed processing of large data sets across clusters of computers"; + license = licenses.asl20; + + longDescription = '' + The Hadoop YARN Container Executor is a native component responsible for managing the lifecycle of containers + on individual nodes in a Hadoop YARN cluster. It launches, monitors, and terminates containers, ensuring that + resources like CPU and memory are allocated according to the policies defined in the ResourceManager. + ''; + + maintainers = with maintainers; [ illustris ]; + platforms = filter (strings.hasSuffix "linux") (attrNames platformAttrs); + }; +}) diff --git a/nixpkgs/pkgs/applications/networking/cluster/hadoop/default.nix b/nixpkgs/pkgs/applications/networking/cluster/hadoop/default.nix index 65512de2031b..d5bae9ad885b 100644 --- a/nixpkgs/pkgs/applications/networking/cluster/hadoop/default.nix +++ b/nixpkgs/pkgs/applications/networking/cluster/hadoop/default.nix @@ -19,6 +19,8 @@ , nixosTests , sparkSupport ? true , spark +, libtirpc +, callPackage }: with lib; @@ -26,40 +28,75 @@ with lib; assert elem stdenv.system [ "x86_64-linux" "x86_64-darwin" "aarch64-linux" "aarch64-darwin" ]; let - common = { pname, platformAttrs, untarDir ? "${pname}-${version}", jdk, openssl ? null, nativeLibs ? [ ], libPatches ? "", tests }: - stdenv.mkDerivation rec { - inherit pname jdk libPatches untarDir openssl; + common = { pname, platformAttrs, jdk, tests }: + stdenv.mkDerivation (finalAttrs: { + inherit pname jdk; version = platformAttrs.${stdenv.system}.version or (throw "Unsupported system: ${stdenv.system}"); src = fetchurl { - url = "mirror://apache/hadoop/common/hadoop-${version}/hadoop-${version}" + optionalString stdenv.isAarch64 "-aarch64" + ".tar.gz"; + url = "mirror://apache/hadoop/common/hadoop-${finalAttrs.version}/hadoop-${finalAttrs.version}" + + optionalString stdenv.isAarch64 "-aarch64" + ".tar.gz"; inherit (platformAttrs.${stdenv.system}) hash; }; doCheck = true; + # Build the container executor binary from source + # InstallPhase is not lazily evaluating containerExecutor for some reason + containerExecutor = if stdenv.isLinux then (callPackage ./containerExecutor.nix { + inherit (finalAttrs) version; + inherit platformAttrs; + }) else ""; + nativeBuildInputs = [ makeWrapper ] - ++ optionals (stdenv.isLinux && (nativeLibs != [ ] || libPatches != "")) [ autoPatchelfHook ]; - buildInputs = [ openssl ] ++ nativeLibs; + ++ optionals stdenv.isLinux [ autoPatchelfHook ]; + buildInputs = optionals stdenv.isLinux [ stdenv.cc.cc.lib openssl protobuf zlib snappy libtirpc ]; installPhase = '' - mkdir -p $out/{lib/${untarDir}/conf,bin,lib} - mv * $out/lib/${untarDir} + mkdir $out + mv * $out/ '' + optionalString stdenv.isLinux '' - # All versions need container-executor, but some versions can't use autoPatchelf because of broken SSL versions - patchelf --set-interpreter ${glibc.out}/lib64/ld-linux-x86-64.so.2 $out/lib/${untarDir}/bin/container-executor + for n in $(find ${finalAttrs.containerExecutor}/bin -type f); do + ln -sf "$n" $out/bin + done + + # these libraries are loaded at runtime by the JVM + ln -s ${getLib cyrus_sasl}/lib/libsasl2.so $out/lib/native/libsasl2.so.2 + ln -s ${getLib openssl}/lib/libcrypto.so $out/lib/native/ + ln -s ${getLib zlib}/lib/libz.so.1 $out/lib/native/ + ln -s ${getLib zstd}/lib/libzstd.so.1 $out/lib/native/ + ln -s ${getLib bzip2}/lib/libbz2.so.1 $out/lib/native/ + ln -s ${getLib snappy}/lib/libsnappy.so.1 $out/lib/native/ + + # libjvm.so is in different paths for java 8 and 11 + # libnativetask.so in hadooop 3 and libhdfs.so in hadoop 2 depend on it + find $out/lib/native/ -name 'libnativetask.so*' -o -name 'libhdfs.so*' | \ + xargs -n1 patchelf --add-rpath $(dirname $(find ${finalAttrs.jdk.home} -name libjvm.so | head -n1)) + + # NixOS/nixpkgs#193370 + # This workaround is needed to use protobuf 3.19 + # hadoop 3.3+ depends on protobuf 3.18, 3.2 depends on 3.8 + find $out/lib/native -name 'libhdfspp.so*' | \ + xargs -r -n1 patchelf --replace-needed libprotobuf.so.${ + if (versionAtLeast finalAttrs.version "3.3") then "18" + else "8" + } libprotobuf.so + + patchelf --replace-needed libcrypto.so.1.1 libcrypto.so \ + $out/lib/native/{libhdfs{pp,}.so*,examples/{pipes-sort,wordcount-nopipe,wordcount-part,wordcount-simple}} + '' + '' - for n in $(find $out/lib/${untarDir}/bin -type f ! -name "*.*"); do - makeWrapper "$n" "$out/bin/$(basename $n)"\ - --set-default JAVA_HOME ${jdk.home}\ - --set-default HADOOP_HOME $out/lib/${untarDir}\ + for n in $(find $out/bin -type f ! -name "*.*"); do + wrapProgram "$n"\ + --set-default JAVA_HOME ${finalAttrs.jdk.home}\ + --set-default HADOOP_HOME $out/\ --run "test -d /etc/hadoop-conf && export HADOOP_CONF_DIR=\''${HADOOP_CONF_DIR-'/etc/hadoop-conf/'}"\ - --set-default HADOOP_CONF_DIR $out/lib/${untarDir}/etc/hadoop/\ + --set-default HADOOP_CONF_DIR $out/etc/hadoop/\ --prefix PATH : "${makeBinPath [ bash coreutils which]}"\ - --prefix JAVA_LIBRARY_PATH : "${makeLibraryPath buildInputs}" + --prefix JAVA_LIBRARY_PATH : "${makeLibraryPath finalAttrs.buildInputs}" done - '' + optionalString sparkSupport '' + '' + (optionalString sparkSupport '' # Add the spark shuffle service jar to YARN - cp ${spark.src}/yarn/spark-${spark.version}-yarn-shuffle.jar $out/lib/${untarDir}/share/hadoop/yarn/ - '' + libPatches; + cp ${spark.src}/yarn/spark-${spark.version}-yarn-shuffle.jar $out/share/hadoop/yarn/ + ''); passthru = { inherit tests; }; @@ -83,7 +120,7 @@ let maintainers = with maintainers; [ illustris ]; platforms = attrNames platformAttrs; } (attrByPath [ stdenv.system "meta" ] {} platformAttrs); - }; + }); in { # Different version of hadoop support different java runtime versions @@ -91,48 +128,29 @@ in hadoop_3_3 = common rec { pname = "hadoop"; platformAttrs = rec { - x86_64-linux = { - version = "3.3.5"; - hash = "sha256-RG4FypL6I6YGF6ixeUbe3kcoGvFQQEFhfLfV9i50JSo="; - }; - x86_64-darwin = x86_64-linux; - aarch64-linux = { - version = "3.3.5"; - hash = "sha256-qcKjbE881isauWBxIv+NY0UFbYit704/Re8Kdl6x1LA="; - }; - aarch64-darwin = aarch64-linux; + x86_64-linux = { + version = "3.3.6"; + hash = "sha256-9RlQWcDUECrap//xf3sqhd+Qa8tuGZSHFjGfmXhkGgQ="; + srcHash = "sha256-4OEsVhBNV9CJ+PN4FgCduUCVA9/el5yezSCZ6ko3+bU="; + }; + x86_64-darwin = x86_64-linux; + aarch64-linux = x86_64-linux // { + hash = "sha256-5Lv2uA72BJEva5v2yncyPe5gKNCNOPNsoHffVt6KXQ0="; + }; + aarch64-darwin = aarch64-linux; }; - untarDir = "${pname}-${platformAttrs.${stdenv.system}.version}"; jdk = jdk11_headless; - inherit openssl; # TODO: Package and add Intel Storage Acceleration Library - nativeLibs = [ stdenv.cc.cc.lib protobuf zlib snappy ]; - libPatches = '' - ln -s ${getLib cyrus_sasl}/lib/libsasl2.so $out/lib/${untarDir}/lib/native/libsasl2.so.2 - ln -s ${getLib openssl}/lib/libcrypto.so $out/lib/${untarDir}/lib/native/ - ln -s ${getLib zlib}/lib/libz.so.1 $out/lib/${untarDir}/lib/native/ - ln -s ${getLib zstd}/lib/libzstd.so.1 $out/lib/${untarDir}/lib/native/ - ln -s ${getLib bzip2}/lib/libbz2.so.1 $out/lib/${untarDir}/lib/native/ - '' + optionalString stdenv.isLinux '' - # libjvm.so for Java >=11 - patchelf --add-rpath ${jdk.home}/lib/server $out/lib/${untarDir}/lib/native/libnativetask.so.1.0.0 - # Java 8 has libjvm.so at a different path - patchelf --add-rpath ${jdk.home}/jre/lib/amd64/server $out/lib/${untarDir}/lib/native/libnativetask.so.1.0.0 - # NixOS/nixpkgs#193370 - # This workaround is needed to use protobuf 3.19 - patchelf --replace-needed libprotobuf.so.18 libprotobuf.so $out/lib/${untarDir}/lib/native/libhdfspp.so - ''; tests = nixosTests.hadoop; }; - hadoop_3_2 = common rec { + hadoop_3_2 = common { pname = "hadoop"; platformAttrs.x86_64-linux = { version = "3.2.4"; hash = "sha256-qt2gpMr+NHuiVR+/zFRzRyRKG725/ZNBIM69z9J9wNw="; + srcHash = "sha256-F9nGD3mZZ1eJf3Ec3AJGE9YBcL/HiagskcdKQhCn/sw="; }; jdk = jdk8_headless; - # not using native libs because of broken openssl_1_0_2 dependency - # can be manually overridden tests = nixosTests.hadoop_3_2; }; hadoop2 = common rec { @@ -140,6 +158,7 @@ in platformAttrs.x86_64-linux = { version = "2.10.2"; hash = "sha256-xhA4zxqIRGNhIeBnJO9dLKf/gx/Bq+uIyyZwsIafEyo="; + srcHash = "sha256-ucxCyXiJo8aL6aNMhZgKEbn8sGKOoMPVREbMGSfSdAI="; }; jdk = jdk8_headless; tests = nixosTests.hadoop2; |