Compare commits
10 Commits
7003c8058a
...
6f00704fa0
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
6f00704fa0 | ||
|
|
d83fdbdf18 | ||
|
|
eec1c40d99 | ||
|
|
0af5dbcc5f | ||
|
|
daa0e99658 | ||
|
|
38d63cf76d | ||
|
|
bf336bb857 | ||
|
|
1db6ed0535 | ||
|
|
7944f9015a | ||
|
|
eddeb27d28 |
14
02-Upgrade-os-maven-plugin-to-1.7.1.patch
Normal file
14
02-Upgrade-os-maven-plugin-to-1.7.1.patch
Normal file
@ -0,0 +1,14 @@
|
|||||||
|
diff --git a/hadoop-project/pom.xml b/hadoop-project/pom.xml
|
||||||
|
index 0b2f6f17..c402535f 100644
|
||||||
|
--- a/hadoop-project/pom.xml
|
||||||
|
+++ b/hadoop-project/pom.xml
|
||||||
|
@@ -151,7 +151,7 @@
|
||||||
|
|
||||||
|
<protobuf-compile.version>3.5.1</protobuf-compile.version>
|
||||||
|
<grpc.version>1.10.0</grpc.version>
|
||||||
|
- <os-maven-plugin.version>1.7.0</os-maven-plugin.version>
|
||||||
|
+ <os-maven-plugin.version>1.7.1</os-maven-plugin.version>
|
||||||
|
|
||||||
|
<!-- define the Java language version used by the compiler -->
|
||||||
|
<javac.version>1.8</javac.version>
|
||||||
|
|
||||||
23
03-Fix-build-on-riscv.patch
Normal file
23
03-Fix-build-on-riscv.patch
Normal file
@ -0,0 +1,23 @@
|
|||||||
|
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-nativetask/src/main/native/src/lib/primitives.h b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-nativetask/src/main/native/src/lib/primitives.h
|
||||||
|
index 000c4b91..70e94487 100644
|
||||||
|
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-nativetask/src/main/native/src/lib/primitives.h
|
||||||
|
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-nativetask/src/main/native/src/lib/primitives.h
|
||||||
|
@@ -99,7 +99,7 @@ inline void simple_memcpy(void * dest, const void * src, size_t len) {
|
||||||
|
inline uint32_t bswap(uint32_t val) {
|
||||||
|
#ifdef __aarch64__
|
||||||
|
__asm__("rev %w[dst], %w[src]" : [dst]"=r"(val) : [src]"r"(val));
|
||||||
|
-#elif defined(__ppc64__)||(__PPC64__)||(__powerpc64__)
|
||||||
|
+#elif defined(__ppc64__)||(__PPC64__)||(__powerpc64__)||(__riscv)
|
||||||
|
return __builtin_bswap32(val);
|
||||||
|
#else
|
||||||
|
__asm__("bswap %0" : "=r" (val) : "0" (val));
|
||||||
|
@@ -110,7 +110,7 @@ inline uint32_t bswap(uint32_t val) {
|
||||||
|
inline uint64_t bswap64(uint64_t val) {
|
||||||
|
#ifdef __aarch64__
|
||||||
|
__asm__("rev %[dst], %[src]" : [dst]"=r"(val) : [src]"r"(val));
|
||||||
|
-#elif defined(__ppc64__)||(__PPC64__)||(__powerpc64__)
|
||||||
|
+#elif defined(__ppc64__)||(__PPC64__)||(__powerpc64__)||(__riscv)
|
||||||
|
return __builtin_bswap64(val);
|
||||||
|
#else
|
||||||
|
#ifdef __X64
|
||||||
|
|
||||||
57
04-Enhance-access-control-for-RunJar.patch
Normal file
57
04-Enhance-access-control-for-RunJar.patch
Normal file
@ -0,0 +1,57 @@
|
|||||||
|
From 7c30e7ffb65f9a58a85b3b556f8c0de04c1b4b20 Mon Sep 17 00:00:00 2001
|
||||||
|
From: He Xiaoqiao <hexiaoqiao@apache.org>
|
||||||
|
Date: Mon, 15 Jan 2024 16:01:08 +0800
|
||||||
|
Subject: [PATCH] HADOOP-19031. Enhance access control for RunJar.
|
||||||
|
|
||||||
|
---
|
||||||
|
.../main/java/org/apache/hadoop/util/RunJar.java | 16 +++++++++-------
|
||||||
|
1 file changed, 9 insertions(+), 7 deletions(-)
|
||||||
|
|
||||||
|
diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/RunJar.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/RunJar.java
|
||||||
|
index c28e69f5..e527f602 100644
|
||||||
|
--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/RunJar.java
|
||||||
|
+++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/RunJar.java
|
||||||
|
@@ -28,10 +28,14 @@ import java.net.MalformedURLException;
|
||||||
|
import java.net.URL;
|
||||||
|
import java.net.URLClassLoader;
|
||||||
|
import java.nio.file.Files;
|
||||||
|
+import java.nio.file.attribute.FileAttribute;
|
||||||
|
+import java.nio.file.attribute.PosixFilePermission;
|
||||||
|
+import java.nio.file.attribute.PosixFilePermissions;
|
||||||
|
import java.util.ArrayList;
|
||||||
|
import java.util.Arrays;
|
||||||
|
import java.util.Enumeration;
|
||||||
|
import java.util.List;
|
||||||
|
+import java.util.Set;
|
||||||
|
import java.util.jar.JarEntry;
|
||||||
|
import java.util.jar.JarFile;
|
||||||
|
import java.util.jar.JarInputStream;
|
||||||
|
@@ -287,20 +291,18 @@ public class RunJar {
|
||||||
|
|
||||||
|
final File workDir;
|
||||||
|
try {
|
||||||
|
- workDir = File.createTempFile("hadoop-unjar", "", tmpDir);
|
||||||
|
- } catch (IOException ioe) {
|
||||||
|
+ FileAttribute<Set<PosixFilePermission>> perms = PosixFilePermissions
|
||||||
|
+ .asFileAttribute(PosixFilePermissions.fromString("rwx------"));
|
||||||
|
+ workDir = Files.createTempDirectory(tmpDir.toPath(), "hadoop-unjar", perms).toFile();
|
||||||
|
+ } catch (IOException | SecurityException e) {
|
||||||
|
// If user has insufficient perms to write to tmpDir, default
|
||||||
|
// "Permission denied" message doesn't specify a filename.
|
||||||
|
System.err.println("Error creating temp dir in java.io.tmpdir "
|
||||||
|
- + tmpDir + " due to " + ioe.getMessage());
|
||||||
|
+ + tmpDir + " due to " + e.getMessage());
|
||||||
|
System.exit(-1);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
- if (!workDir.delete()) {
|
||||||
|
- System.err.println("Delete failed for " + workDir);
|
||||||
|
- System.exit(-1);
|
||||||
|
- }
|
||||||
|
ensureDirectory(workDir);
|
||||||
|
|
||||||
|
ShutdownHookManager.get().addShutdownHook(
|
||||||
|
--
|
||||||
|
2.43.0
|
||||||
|
|
||||||
94
hadoop.spec
94
hadoop.spec
@ -9,9 +9,11 @@
|
|||||||
%global __requires_exclude_from ^%{_libdir}/%{name}/libhadoop.so$
|
%global __requires_exclude_from ^%{_libdir}/%{name}/libhadoop.so$
|
||||||
%global __provides_exclude_from ^%{_libdir}/%{name}/.*$
|
%global __provides_exclude_from ^%{_libdir}/%{name}/.*$
|
||||||
%define _binaries_in_noarch_packages_terminate_build 0
|
%define _binaries_in_noarch_packages_terminate_build 0
|
||||||
|
%define huawei_repo https://repo.huaweicloud.com/repository
|
||||||
|
|
||||||
Name: hadoop
|
Name: hadoop
|
||||||
Version: 3.3.6
|
Version: 3.3.6
|
||||||
Release: 1
|
Release: 6
|
||||||
Summary: A software platform for processing vast amounts of data
|
Summary: A software platform for processing vast amounts of data
|
||||||
# The BSD license file is missing
|
# The BSD license file is missing
|
||||||
# https://issues.apache.org/jira/browse/HADOOP-9849
|
# https://issues.apache.org/jira/browse/HADOOP-9849
|
||||||
@ -34,12 +36,16 @@ Source13: %{name}-yarn-site.xml
|
|||||||
Source14: yarn-v1.22.5.tar.gz
|
Source14: yarn-v1.22.5.tar.gz
|
||||||
Source15: node-12.22.1-linux-x64.tar.gz
|
Source15: node-12.22.1-linux-x64.tar.gz
|
||||||
Source16: node-v12.22.1-linux-arm64.tar.gz
|
Source16: node-v12.22.1-linux-arm64.tar.gz
|
||||||
Source17: settings.xml
|
Source17: protoc-3.7.1-linux-riscv64.exe
|
||||||
|
Source18: protoc-gen-grpc-java-1.26.0-linux-riscv_64.exe
|
||||||
|
|
||||||
Patch0: 01-lock-triple-beam-version-to-1.3.0.patch
|
Patch0: 01-lock-triple-beam-version-to-1.3.0.patch
|
||||||
|
Patch1: 02-Upgrade-os-maven-plugin-to-1.7.1.patch
|
||||||
|
Patch2: 03-Fix-build-on-riscv.patch
|
||||||
|
Patch3: 04-Enhance-access-control-for-RunJar.patch
|
||||||
|
|
||||||
BuildRoot: %{_tmppath}/%{name}-%{version}-%{release}-root
|
BuildRoot: %{_tmppath}/%{name}-%{version}-%{release}-root
|
||||||
BuildRequires: java-1.8.0-openjdk-devel maven hostname maven-local tomcat cmake snappy openssl-devel
|
BuildRequires: java-1.8.0-openjdk-devel maven hostname maven-local tomcat cmake snappy openssl-devel
|
||||||
BuildRequires: cyrus-sasl-devel chrpath systemd protobuf2-compiler protobuf2-devel protobuf2-java protobuf2
|
BuildRequires: cyrus-sasl-devel chrpath systemd protobuf2-compiler protobuf2-devel protobuf2-java protobuf2
|
||||||
BuildRequires: leveldbjni leveldb-java hawtjni-runtime gcc-c++
|
BuildRequires: leveldbjni leveldb-java hawtjni-runtime gcc-c++
|
||||||
BuildRequires: npm chrpath
|
BuildRequires: npm chrpath
|
||||||
@ -76,6 +82,7 @@ Obsoletes: %{name}-javadoc < 2.4.1-22%{?dist}
|
|||||||
Requires: apache-zookeeper
|
Requires: apache-zookeeper
|
||||||
Requires: leveldb
|
Requires: leveldb
|
||||||
Requires: protobuf2-java
|
Requires: protobuf2-java
|
||||||
|
Conflicts: hadoop-3.1-client
|
||||||
|
|
||||||
%description common
|
%description common
|
||||||
Apache Hadoop is a framework that allows for the distributed processing of
|
Apache Hadoop is a framework that allows for the distributed processing of
|
||||||
@ -89,6 +96,7 @@ Hadoop modules.
|
|||||||
%package common-native
|
%package common-native
|
||||||
Summary: The native Apache Hadoop library file
|
Summary: The native Apache Hadoop library file
|
||||||
Requires: %{name}-common = %{version}-%{release}
|
Requires: %{name}-common = %{version}-%{release}
|
||||||
|
Conflicts: hadoop-3.1-common
|
||||||
|
|
||||||
%description common-native
|
%description common-native
|
||||||
Apache Hadoop is a framework that allows for the distributed processing of
|
Apache Hadoop is a framework that allows for the distributed processing of
|
||||||
@ -101,6 +109,7 @@ This package contains the native-hadoop library
|
|||||||
%package devel
|
%package devel
|
||||||
Summary: Headers for Apache Hadoop
|
Summary: Headers for Apache Hadoop
|
||||||
Requires: libhdfs%{?_isa} = %{version}-%{release}
|
Requires: libhdfs%{?_isa} = %{version}-%{release}
|
||||||
|
Conflicts: hadoop-3.1-common-native
|
||||||
|
|
||||||
%description devel
|
%description devel
|
||||||
Header files for Apache Hadoop's hdfs library and other utilities
|
Header files for Apache Hadoop's hdfs library and other utilities
|
||||||
@ -113,6 +122,7 @@ Requires: %{name}-common = %{version}-%{release}
|
|||||||
Requires(post): systemd
|
Requires(post): systemd
|
||||||
Requires(preun): systemd
|
Requires(preun): systemd
|
||||||
Requires(postun): systemd
|
Requires(postun): systemd
|
||||||
|
Conflicts: hadoop-3.1-hdfs
|
||||||
|
|
||||||
%description hdfs
|
%description hdfs
|
||||||
Apache Hadoop is a framework that allows for the distributed processing of
|
Apache Hadoop is a framework that allows for the distributed processing of
|
||||||
@ -136,6 +146,7 @@ Requires: tcnative
|
|||||||
Requires(post): systemd
|
Requires(post): systemd
|
||||||
Requires(preun): systemd
|
Requires(preun): systemd
|
||||||
Requires(postun): systemd
|
Requires(postun): systemd
|
||||||
|
Conflicts: hadoop-3.1-httpfs
|
||||||
|
|
||||||
%description httpfs
|
%description httpfs
|
||||||
Apache Hadoop is a framework that allows for the distributed processing of
|
Apache Hadoop is a framework that allows for the distributed processing of
|
||||||
@ -150,6 +161,7 @@ the complete FileSystem/FileContext interface in HDFS.
|
|||||||
Summary: The Apache Hadoop Filesystem Library
|
Summary: The Apache Hadoop Filesystem Library
|
||||||
Requires: %{name}-hdfs = %{version}-%{release}
|
Requires: %{name}-hdfs = %{version}-%{release}
|
||||||
Requires: lzo
|
Requires: lzo
|
||||||
|
Conflicts: hadoop-3.1-libhdfs
|
||||||
|
|
||||||
%description -n libhdfs
|
%description -n libhdfs
|
||||||
Apache Hadoop is a framework that allows for the distributed processing of
|
Apache Hadoop is a framework that allows for the distributed processing of
|
||||||
@ -167,6 +179,7 @@ Requires: %{name}-mapreduce-examples = %{version}-%{release}
|
|||||||
Requires(post): systemd
|
Requires(post): systemd
|
||||||
Requires(preun): systemd
|
Requires(preun): systemd
|
||||||
Requires(postun): systemd
|
Requires(postun): systemd
|
||||||
|
Conflicts: hadoop-3.1-mapreduce
|
||||||
|
|
||||||
%description mapreduce
|
%description mapreduce
|
||||||
Apache Hadoop is a framework that allows for the distributed processing of
|
Apache Hadoop is a framework that allows for the distributed processing of
|
||||||
@ -180,6 +193,7 @@ This package provides Apache Hadoop MapReduce (MRv2).
|
|||||||
Summary: Apache Hadoop MapReduce (MRv2) examples
|
Summary: Apache Hadoop MapReduce (MRv2) examples
|
||||||
BuildArch: noarch
|
BuildArch: noarch
|
||||||
Requires: hsqldb
|
Requires: hsqldb
|
||||||
|
Conflicts: hadoop-3.1-mapreduce-examples
|
||||||
|
|
||||||
%description mapreduce-examples
|
%description mapreduce-examples
|
||||||
This package contains mapreduce examples.
|
This package contains mapreduce examples.
|
||||||
@ -188,6 +202,7 @@ This package contains mapreduce examples.
|
|||||||
Summary: Apache Hadoop maven plugin
|
Summary: Apache Hadoop maven plugin
|
||||||
BuildArch: noarch
|
BuildArch: noarch
|
||||||
Requires: maven
|
Requires: maven
|
||||||
|
Conflicts: hadoop-3.1-maven-plugin
|
||||||
|
|
||||||
%description maven-plugin
|
%description maven-plugin
|
||||||
The Apache Hadoop maven plugin
|
The Apache Hadoop maven plugin
|
||||||
@ -199,6 +214,7 @@ Requires: %{name}-common = %{version}-%{release}
|
|||||||
Requires: %{name}-hdfs = %{version}-%{release}
|
Requires: %{name}-hdfs = %{version}-%{release}
|
||||||
Requires: %{name}-mapreduce = %{version}-%{release}
|
Requires: %{name}-mapreduce = %{version}-%{release}
|
||||||
Requires: %{name}-yarn = %{version}-%{release}
|
Requires: %{name}-yarn = %{version}-%{release}
|
||||||
|
Conflicts: hadoop-3.1-tests
|
||||||
|
|
||||||
%description tests
|
%description tests
|
||||||
Apache Hadoop is a framework that allows for the distributed processing of
|
Apache Hadoop is a framework that allows for the distributed processing of
|
||||||
@ -220,6 +236,7 @@ Requires: leveldbjni
|
|||||||
Requires(post): systemd
|
Requires(post): systemd
|
||||||
Requires(preun): systemd
|
Requires(preun): systemd
|
||||||
Requires(postun): systemd
|
Requires(postun): systemd
|
||||||
|
Conflicts: hadoop-3.1-yarn nodejs-yarn
|
||||||
|
|
||||||
%description yarn
|
%description yarn
|
||||||
Apache Hadoop is a framework that allows for the distributed processing of
|
Apache Hadoop is a framework that allows for the distributed processing of
|
||||||
@ -232,6 +249,7 @@ This package contains Apache Hadoop YARN.
|
|||||||
%package yarn-security
|
%package yarn-security
|
||||||
Summary: The ability to run Apache Hadoop YARN in secure mode
|
Summary: The ability to run Apache Hadoop YARN in secure mode
|
||||||
Requires: %{name}-yarn = %{version}-%{release}
|
Requires: %{name}-yarn = %{version}-%{release}
|
||||||
|
Conflicts: hadoop-3.1-yarn-security
|
||||||
|
|
||||||
%description yarn-security
|
%description yarn-security
|
||||||
Apache Hadoop is a framework that allows for the distributed processing of
|
Apache Hadoop is a framework that allows for the distributed processing of
|
||||||
@ -243,27 +261,30 @@ This package contains files needed to run Apache Hadoop YARN in secure mode.
|
|||||||
|
|
||||||
%prep
|
%prep
|
||||||
%autosetup -p1 -n %{name}-%{version}-src
|
%autosetup -p1 -n %{name}-%{version}-src
|
||||||
cp %{SOURCE17} ./
|
%ifarch riscv64
|
||||||
sed -i "s,@HOME@,${HOME},g" settings.xml
|
mvn install:install-file -DgroupId=com.google.protobuf -DartifactId=protoc -Dversion=2.5.0 -Dclassifier=linux-riscv64 -Dpackaging=exe -Dfile=/usr/bin/protoc
|
||||||
mvn install:install-file -DgroupId=org.fusesource.leveldbjni -DartifactId=leveldbjni-all -Dversion=1.8 -Dpackaging=jar -Dfile=/usr/lib/java/leveldbjni-all.jar -s settings.xml
|
mvn install:install-file -DgroupId=com.google.protobuf -DartifactId=protoc -Dversion=3.7.1 -Dclassifier=linux-riscv64 -Dpackaging=exe -Dfile=%{SOURCE17}
|
||||||
mvn install:install-file -DgroupId=org.fusesource.leveldbjni -DartifactId=leveldbjni -Dversion=1.8 -Dpackaging=jar -Dfile=/usr/lib/java/leveldbjni/leveldbjni.jar -s settings.xml
|
mvn install:install-file -DgroupId=io.grpc -DartifactId=protoc-gen-grpc-java -Dversion=1.26.0 -Dclassifier=linux-riscv64 -Dpackaging=exe -Dfile=%{SOURCE18}
|
||||||
mvn install:install-file -DgroupId=org.iq80.leveldb -DartifactId=leveldb-api -Dversion=0.7 -Dpackaging=jar -Dfile=/usr/share/java/leveldb-java/leveldb-api.jar -s settings.xml
|
%endif
|
||||||
mvn install:install-file -DgroupId=org.iq80.leveldb -DartifactId=leveldb-benchmark -Dversion=0.7 -Dpackaging=jar -Dfile=/usr/share/java/leveldb-java/leveldb-benchmark.jar -s settings.xml
|
mvn install:install-file -DgroupId=org.fusesource.leveldbjni -DartifactId=leveldbjni-all -Dversion=1.8 -Dpackaging=jar -Dfile=/usr/lib/java/leveldbjni-all.jar
|
||||||
mvn install:install-file -DgroupId=org.iq80.leveldb -DartifactId=leveldb -Dversion=0.7 -Dpackaging=jar -Dfile=/usr/share/java/leveldb-java/leveldb.jar -s settings.xml
|
mvn install:install-file -DgroupId=org.fusesource.leveldbjni -DartifactId=leveldbjni -Dversion=1.8 -Dpackaging=jar -Dfile=/usr/lib/java/leveldbjni/leveldbjni.jar
|
||||||
mvn install:install-file -DgroupId=orn.fusesource.hawtjni -DartifactId=hawtjni-runtime -Dversion=1.16 -Dpackaging=jar -Dfile=/usr/lib/java/hawtjni/hawtjni-runtime.jar -s settings.xml
|
mvn install:install-file -DgroupId=org.iq80.leveldb -DartifactId=leveldb-api -Dversion=0.7 -Dpackaging=jar -Dfile=/usr/share/java/leveldb-java/leveldb-api.jar
|
||||||
|
mvn install:install-file -DgroupId=org.iq80.leveldb -DartifactId=leveldb-benchmark -Dversion=0.7 -Dpackaging=jar -Dfile=/usr/share/java/leveldb-java/leveldb-benchmark.jar
|
||||||
|
mvn install:install-file -DgroupId=org.iq80.leveldb -DartifactId=leveldb -Dversion=0.7 -Dpackaging=jar -Dfile=/usr/share/java/leveldb-java/leveldb.jar
|
||||||
|
mvn install:install-file -DgroupId=orn.fusesource.hawtjni -DartifactId=hawtjni-runtime -Dversion=1.16 -Dpackaging=jar -Dfile=/usr/lib/java/hawtjni/hawtjni-runtime.jar
|
||||||
|
|
||||||
mkdir -p ${HOME}/repository/com/github/eirslett/node/12.22.1/
|
mkdir -p ${HOME}/.m2/repository/com/github/eirslett/node/12.22.1/
|
||||||
cp %{SOURCE15} ${HOME}/repository/com/github/eirslett/node/12.22.1/
|
cp %{SOURCE15} ${HOME}/.m2/repository/com/github/eirslett/node/12.22.1/
|
||||||
cp %{SOURCE16} ${HOME}/repository/com/github/eirslett/node/12.22.1/
|
cp %{SOURCE16} ${HOME}/.m2/repository/com/github/eirslett/node/12.22.1/
|
||||||
mv ${HOME}/repository/com/github/eirslett/node/12.22.1/node-v12.22.1-linux-arm64.tar.gz ${HOME}/repository/com/github/eirslett/node/12.22.1/node-12.22.1-linux-arm64.tar.gz
|
mv ${HOME}/.m2/repository/com/github/eirslett/node/12.22.1/node-v12.22.1-linux-arm64.tar.gz ${HOME}/.m2/repository/com/github/eirslett/node/12.22.1/node-12.22.1-linux-arm64.tar.gz
|
||||||
mkdir -p ${HOME}/repository/com/github/eirslett/yarn/1.22.5/
|
mkdir -p ${HOME}/.m2/repository/com/github/eirslett/yarn/1.22.5/
|
||||||
cp %{SOURCE14} ${HOME}/repository/com/github/eirslett/yarn/1.22.5/
|
cp %{SOURCE14} ${HOME}/.m2/repository/com/github/eirslett/yarn/1.22.5/
|
||||||
mv ${HOME}/repository/com/github/eirslett/yarn/1.22.5/yarn-v1.22.5.tar.gz ${HOME}/repository/com/github/eirslett/yarn/1.22.5/yarn-1.22.5.tar.gz
|
mv ${HOME}/.m2/repository/com/github/eirslett/yarn/1.22.5/yarn-v1.22.5.tar.gz ${HOME}/.m2/repository/com/github/eirslett/yarn/1.22.5/yarn-1.22.5.tar.gz
|
||||||
tar -xzvf ${HOME}/repository/com/github/eirslett/yarn/1.22.5/yarn-1.22.5.tar.gz -C ${HOME}/repository/com/github/eirslett/yarn/1.22.5/
|
tar -xzvf ${HOME}/.m2/repository/com/github/eirslett/yarn/1.22.5/yarn-1.22.5.tar.gz -C ${HOME}/.m2/repository/com/github/eirslett/yarn/1.22.5/
|
||||||
npm config set registry https://repo.huaweicloud.com/repository/npm/
|
npm config set registry %{huawei_repo}/npm/
|
||||||
npm cache clean -f
|
npm cache clean -f
|
||||||
${HOME}/repository/com/github/eirslett/yarn/1.22.5/yarn-v1.22.5/bin/yarn config set registry https://repo.huaweicloud.com/repository/npm/ -g
|
${HOME}/.m2/repository/com/github/eirslett/yarn/1.22.5/yarn-v1.22.5/bin/yarn config set registry %{huawei_repo}/npm/ -g
|
||||||
${HOME}/repository/com/github/eirslett/yarn/1.22.5/yarn-v1.22.5/bin/yarn config set ignore-engines true
|
${HOME}/.m2/repository/com/github/eirslett/yarn/1.22.5/yarn-v1.22.5/bin/yarn config set ignore-engines true
|
||||||
|
|
||||||
%pom_add_dep org.iq80.leveldb:leveldb-api:0.7 hadoop-hdfs-project/hadoop-hdfs
|
%pom_add_dep org.iq80.leveldb:leveldb-api:0.7 hadoop-hdfs-project/hadoop-hdfs
|
||||||
%pom_add_dep org.iq80.leveldb:leveldb-api:0.7 hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy
|
%pom_add_dep org.iq80.leveldb:leveldb-api:0.7 hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy
|
||||||
@ -277,6 +298,9 @@ ${HOME}/repository/com/github/eirslett/yarn/1.22.5/yarn-v1.22.5/bin/yarn config
|
|||||||
%pom_disable_module hadoop-azure hadoop-tools
|
%pom_disable_module hadoop-azure hadoop-tools
|
||||||
%pom_disable_module hadoop-yarn-server-timelineservice-hbase-tests hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/pom.xml
|
%pom_disable_module hadoop-yarn-server-timelineservice-hbase-tests hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/pom.xml
|
||||||
|
|
||||||
|
%ifarch riscv64
|
||||||
|
%pom_disable_module hadoop-yarn-applications-catalog-webapp hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-applications-catalog/pom.xml
|
||||||
|
%endif
|
||||||
|
|
||||||
# War files we don't want
|
# War files we don't want
|
||||||
%mvn_package :%{name}-auth-examples __noinstall
|
%mvn_package :%{name}-auth-examples __noinstall
|
||||||
@ -313,7 +337,12 @@ ${HOME}/repository/com/github/eirslett/yarn/1.22.5/yarn-v1.22.5/bin/yarn config
|
|||||||
%mvn_file :%{name}-common::tests: %{name}/%{name}-common
|
%mvn_file :%{name}-common::tests: %{name}/%{name}-common
|
||||||
|
|
||||||
%build
|
%build
|
||||||
mvn clean -Dsnappy.lib=/usr/lib64 -Dbundle.snappy -Dcontainer-executor.conf.dir=%{_sysconfdir}/%{name} -Pdist,native -DskipTests -DskipIT -Dmaven.javadoc.skip=true package -s settings.xml
|
%ifarch riscv64
|
||||||
|
export MAVEN_OPTS="-Xms2048M -Xmx8000M"
|
||||||
|
function keepalive() { while true; do date; sleep 3600; done }
|
||||||
|
keepalive &
|
||||||
|
%endif
|
||||||
|
mvn clean -Dsnappy.lib=/usr/lib64 -Dbundle.snappy -Dcontainer-executor.conf.dir=%{_sysconfdir}/%{name} -Pdist,native -DskipTests -DskipIT -Dmaven.javadoc.skip=true package
|
||||||
|
|
||||||
%install
|
%install
|
||||||
# Copy all jar files except those generated by the build
|
# Copy all jar files except those generated by the build
|
||||||
@ -1133,6 +1162,23 @@ fi
|
|||||||
%config(noreplace) %{_sysconfdir}/%{name}/container-executor.cfg
|
%config(noreplace) %{_sysconfdir}/%{name}/container-executor.cfg
|
||||||
|
|
||||||
%changelog
|
%changelog
|
||||||
|
* Fri Dec 13 2024 xuduo <fs-yao@126.com> - 3.3.6-6
|
||||||
|
- Add huawei repository
|
||||||
|
|
||||||
|
* Mon Nov 11 2024 xiaochn <xiaochuannan@inspur.com> - 3.3.6-5
|
||||||
|
- Remove npm repo in spec file and use unified configuration of the build environment
|
||||||
|
|
||||||
|
* Thu Sep 26 2024 Deyuan Fan <fandeyuan@kylinos.cn> - 3.3.6-4
|
||||||
|
- fix CVE-2024-23454
|
||||||
|
|
||||||
|
* Sun May 12 2024 Dingli Zhang <dingli@iscas.ac.cn> - 3.3.6-3
|
||||||
|
- Fix build on riscv64
|
||||||
|
- Upgrade os-maven-plugin to 1.7.1
|
||||||
|
- Remove settings.xml and use ${HOME}/.m2/repository as maven repo
|
||||||
|
|
||||||
|
* Fri Dec 15 2023 xiexing <xiexing4@hisilicon.com> - 3.3.6-2
|
||||||
|
- add conflicts to hadoop spec
|
||||||
|
|
||||||
* Mon Nov 27 2023 wenweijian wenweijian2@huawei.com - 3.3.6-1
|
* Mon Nov 27 2023 wenweijian wenweijian2@huawei.com - 3.3.6-1
|
||||||
- fix cve CVE-2023-26031
|
- fix cve CVE-2023-26031
|
||||||
|
|
||||||
|
|||||||
BIN
protoc-3.7.1-linux-riscv64.exe
Normal file
BIN
protoc-3.7.1-linux-riscv64.exe
Normal file
Binary file not shown.
BIN
protoc-gen-grpc-java-1.26.0-linux-riscv_64.exe
Normal file
BIN
protoc-gen-grpc-java-1.26.0-linux-riscv_64.exe
Normal file
Binary file not shown.
22
settings.xml
22
settings.xml
@ -1,22 +0,0 @@
|
|||||||
<?xml version="1.0" encoding="UTF-8"?>
|
|
||||||
<settings xmlns="http://maven.apache.org/SETTINGS/1.0.0"
|
|
||||||
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
|
|
||||||
xsi:schemaLocation="http://maven.apache.org/SETTINGS/1.0.0 http://maven.apache.org/xsd/settings-1.0.0.xsd">
|
|
||||||
|
|
||||||
<localRepository>@HOME@/repository/</localRepository>
|
|
||||||
<pluginGroups>
|
|
||||||
</pluginGroups>
|
|
||||||
<proxies>
|
|
||||||
</proxies>
|
|
||||||
<servers>
|
|
||||||
</servers>
|
|
||||||
<mirrors>
|
|
||||||
<mirror>
|
|
||||||
<id>huaweicloud</id>
|
|
||||||
<mirrorOf>*</mirrorOf>
|
|
||||||
<url>https://repo.huaweicloud.com/repository/maven/</url>
|
|
||||||
</mirror>
|
|
||||||
</mirrors>
|
|
||||||
<profiles>
|
|
||||||
</profiles>
|
|
||||||
</settings>
|
|
||||||
Loading…
x
Reference in New Issue
Block a user