[hadoop/f20] Build 2.2.0-2

rrati rrati at fedoraproject.org
Tue Dec 3 18:45:14 UTC 2013


commit 25f0554dabbc30b79ee7d262d2d13dd02402725a
Author: Robert Rati <rrati at redhat.com>
Date:   Tue Dec 3 13:00:52 2013 -0500

    Build 2.2.0-2

 hadoop-fedora-integration.patch |  191 +++++++++++++++++--------------
 hadoop.spec                     |  242 +++++++++++++++++++++------------------
 2 files changed, 237 insertions(+), 196 deletions(-)
---
diff --git a/hadoop-fedora-integration.patch b/hadoop-fedora-integration.patch
index afbb000..91c4cd1 100644
--- a/hadoop-fedora-integration.patch
+++ b/hadoop-fedora-integration.patch
@@ -161,7 +161,7 @@ index 6059d8c..bba1a00 100644
      System.out.println("Running embedded servlet container at: http://" + host + ":" + port);
    }
 diff --git a/hadoop-common-project/hadoop-common/pom.xml b/hadoop-common-project/hadoop-common/pom.xml
-index 89691c6..37e73b5 100644
+index 89691c6..82458e7 100644
 --- a/hadoop-common-project/hadoop-common/pom.xml
 +++ b/hadoop-common-project/hadoop-common/pom.xml
 @@ -53,7 +53,7 @@
@@ -220,7 +220,7 @@ index 89691c6..37e73b5 100644
  
      <dependency>
        <groupId>com.sun.jersey</groupId>
-@@ -113,21 +136,26 @@
+@@ -113,21 +136,30 @@
        <artifactId>jersey-server</artifactId>
        <scope>compile</scope>
      </dependency>
@@ -229,41 +229,43 @@ index 89691c6..37e73b5 100644
 +      <artifactId>jersey-servlet</artifactId>
 +      <scope>compile</scope>
 +    </dependency>
++    <dependency>
++        <groupId>org.glassfish.web</groupId>
++        <artifactId>javax.servlet.jsp</artifactId>
++    </dependency>
  
      <dependency>
 -      <groupId>tomcat</groupId>
 -      <artifactId>jasper-compiler</artifactId>
--      <scope>runtime</scope>
-+      <groupId>org.glassfish.web</groupId>
-+      <artifactId>javax.servlet.jsp</artifactId>
-+      <version>2.2.6</version>
++      <groupId>org.apache.tomcat</groupId>
++      <artifactId>tomcat-jasper</artifactId>
+       <scope>runtime</scope>
      </dependency>
      <dependency>
 -      <groupId>tomcat</groupId>
 -      <artifactId>jasper-runtime</artifactId>
 -      <scope>runtime</scope>
-+      <groupId>org.apache.tomcat</groupId>
-+      <artifactId>tomcat-servlet-api</artifactId>
-+      <version>7.0.37</version>
++        <groupId>org.apache.tomcat</groupId>
++        <artifactId>tomcat-servlet-api</artifactId>
++        <version>7.0.37</version>
      </dependency>
      <dependency>
 -      <groupId>javax.servlet.jsp</groupId>
 -      <artifactId>jsp-api</artifactId>
 -      <scope>runtime</scope>
-+      <groupId>org.apache.tomcat</groupId>
-+      <artifactId>tomcat-el-api</artifactId>
-+      <version>7.0.37</version>
++        <groupId>org.apache.tomcat</groupId>
++        <artifactId>tomcat-el-api</artifactId>
++        <version>7.0.37</version>
      </dependency>
      <dependency>
        <groupId>commons-el</groupId>
-@@ -213,6 +241,11 @@
+@@ -213,6 +245,10 @@
        <groupId>com.jcraft</groupId>
        <artifactId>jsch</artifactId>
      </dependency>
 +    <dependency>
 +      <groupId>com.google.code.findbugs</groupId>
 +      <artifactId>jsr305</artifactId>
-+      <version>1.3.9</version>
 +    </dependency>
  
      <dependency>
@@ -1596,7 +1598,7 @@ index 4442281..44cf67e 100644
        throw new RuntimeException("It should never happen, " + ex.getMessage(), ex);
      }
 diff --git a/hadoop-hdfs-project/hadoop-hdfs-nfs/pom.xml b/hadoop-hdfs-project/hadoop-hdfs-nfs/pom.xml
-index 13872c3..c2ceab7 100644
+index 13872c3..38d2a14 100644
 --- a/hadoop-hdfs-project/hadoop-hdfs-nfs/pom.xml
 +++ b/hadoop-hdfs-project/hadoop-hdfs-nfs/pom.xml
 @@ -85,12 +85,12 @@ http://maven.apache.org/xsd/maven-4.0.0.xsd">
@@ -1615,20 +1617,30 @@ index 13872c3..c2ceab7 100644
        <artifactId>jetty-util</artifactId>
        <scope>compile</scope>
      </dependency>
-@@ -180,11 +180,6 @@ http://maven.apache.org/xsd/maven-4.0.0.xsd">
+@@ -135,8 +135,8 @@ http://maven.apache.org/xsd/maven-4.0.0.xsd">
+       <scope>compile</scope>
+     </dependency>
+     <dependency>
+-      <groupId>javax.servlet.jsp</groupId>
+-      <artifactId>jsp-api</artifactId>
++      <groupId>org.glassfish.web</groupId>
++      <artifactId>javax.servlet.jsp</artifactId>
+       <scope>compile</scope>
+     </dependency>
+     <dependency>
+@@ -180,8 +180,8 @@ http://maven.apache.org/xsd/maven-4.0.0.xsd">
        <scope>compile</scope>
      </dependency>
      <dependency>
 -      <groupId>tomcat</groupId>
 -      <artifactId>jasper-runtime</artifactId>
--      <scope>compile</scope>
--    </dependency>
--    <dependency>
-       <groupId>xmlenc</groupId>
-       <artifactId>xmlenc</artifactId>
++      <groupId>org.apache.tomcat</groupId>
++      <artifactId>tomcat-jasper</artifactId>
        <scope>compile</scope>
+     </dependency>
+     <dependency>
 diff --git a/hadoop-hdfs-project/hadoop-hdfs/pom.xml b/hadoop-hdfs-project/hadoop-hdfs/pom.xml
-index f8f2918..fd43842 100644
+index f8f2918..b62015e 100644
 --- a/hadoop-hdfs-project/hadoop-hdfs/pom.xml
 +++ b/hadoop-hdfs-project/hadoop-hdfs/pom.xml
 @@ -71,12 +71,12 @@ http://maven.apache.org/xsd/maven-4.0.0.xsd">
@@ -1671,19 +1683,18 @@ index f8f2918..fd43842 100644
        <groupId>junit</groupId>
        <artifactId>junit</artifactId>
        <scope>test</scope>
-@@ -166,11 +156,6 @@ http://maven.apache.org/xsd/maven-4.0.0.xsd">
+@@ -166,8 +156,8 @@ http://maven.apache.org/xsd/maven-4.0.0.xsd">
        <scope>compile</scope>
      </dependency>
      <dependency>
 -      <groupId>tomcat</groupId>
 -      <artifactId>jasper-runtime</artifactId>
--      <scope>compile</scope>
--    </dependency>
--    <dependency>
-       <groupId>xmlenc</groupId>
-       <artifactId>xmlenc</artifactId>
++      <groupId>org.apache.tomcat</groupId>
++      <artifactId>tomcat-jasper</artifactId>
        <scope>compile</scope>
-@@ -192,101 +177,90 @@ http://maven.apache.org/xsd/maven-4.0.0.xsd">
+     </dependency>
+     <dependency>
+@@ -192,101 +182,77 @@ http://maven.apache.org/xsd/maven-4.0.0.xsd">
          </configuration>
        </plugin>
        <plugin>
@@ -1794,42 +1805,33 @@ index f8f2918..fd43842 100644
            </execution>
          </executions>
          <dependencies>
-+          <!-- volatile mix of glassfish 2.2 jspc impl & Tomcat's servlet 3.0 api
-+               EXERCISE CAUTION when messing with these
-+          -->
-+          <dependency>
-+            <groupId>org.glassfish.web</groupId>
-+            <artifactId>javax.servlet.jsp</artifactId>
-+            <version>2.2.5</version>
-+            <scope>runtime</scope>
-+          </dependency>
            <dependency>
 -            <groupId>org.codehaus.mojo.jspc</groupId>
 -            <artifactId>jspc-compiler-tomcat5</artifactId>
 -            <version>2.0-alpha-3</version>
-+            <groupId>org.apache.tomcat</groupId>
-+            <artifactId>tomcat-servlet-api</artifactId>
-+            <version>7.0.37</version>
-           </dependency>
-           <dependency>
+-          </dependency>
+-          <dependency>
 -            <groupId>org.slf4j</groupId>
 -            <artifactId>slf4j-log4j12</artifactId>
 -            <version>1.4.1</version>
-+            <groupId>org.apache.tomcat</groupId>
-+            <artifactId>tomcat-el-api</artifactId>
-+            <version>7.0.37</version>
-           </dependency>
-           <dependency>
+-          </dependency>
+-          <dependency>
 -            <groupId>org.slf4j</groupId>
 -            <artifactId>jcl104-over-slf4j</artifactId>
 -            <version>1.4.1</version>
 +            <groupId>org.apache.hadoop</groupId>
 +            <artifactId>hadoop-common</artifactId>
 +            <version>${project.version}</version>
++            <exclusions>
++              <exclusion>
++                <groupId>javax.servlet</groupId>
++                <artifactId>servlet-api</artifactId>
++               </exclusion>
++            </exclusions>
            </dependency>
          </dependencies>
        </plugin>
-@@ -329,7 +303,7 @@ http://maven.apache.org/xsd/maven-4.0.0.xsd">
+@@ -329,7 +295,7 @@ http://maven.apache.org/xsd/maven-4.0.0.xsd">
            </execution>
            <execution>
              <id>create-web-xmls</id>
@@ -2463,7 +2465,7 @@ index 11640d7..6290e72 100644
        </exclusions>
      </dependency>
 diff --git a/hadoop-project/pom.xml b/hadoop-project/pom.xml
-index b7a7456..abdf900 100644
+index b7a7456..0bee9b1 100644
 --- a/hadoop-project/pom.xml
 +++ b/hadoop-project/pom.xml
 @@ -63,7 +63,7 @@
@@ -2516,7 +2518,20 @@ index b7a7456..abdf900 100644
        </dependency>
  
        <dependency>
-@@ -404,12 +398,23 @@
+@@ -368,6 +362,12 @@
+       </dependency>
+ 
+       <dependency>
++        <groupId>org.glassfish.web</groupId>
++        <artifactId>javax.servlet.jsp</artifactId>
++        <version>2.2.6</version>
++      </dependency>
++
++      <dependency>
+         <groupId>org.codehaus.plexus</groupId>
+         <artifactId>plexus-utils</artifactId>
+         <version>2.0.5</version>
+@@ -404,12 +404,23 @@
          <artifactId>jersey-server</artifactId>
          <version>${jersey.version}</version>
        </dependency>
@@ -2540,54 +2555,50 @@ index b7a7456..abdf900 100644
  
        <dependency>
          <groupId>com.google.inject.extensions</groupId>
-@@ -448,34 +453,22 @@
+@@ -448,14 +459,9 @@
        </dependency>
  
        <dependency>
 -        <groupId>org.mortbay.jetty</groupId>
 -        <artifactId>jetty-servlet-tester</artifactId>
 -        <version>6.1.26</version>
-+        <groupId>org.eclipse.jetty</groupId>
-+        <artifactId>test-jetty-servlet</artifactId>
-+        <version>9.0.4.v20130625</version>
-       </dependency>
-+      <!-- volatile mix of glassfish 2.2 jspc & Tomcat's servlet 3.0 api
-+           EXERCISE CAUTION when messing with these
-+      -->
-       <dependency>
--        <groupId>tomcat</groupId>
--        <artifactId>jasper-compiler</artifactId>
--        <version>5.5.23</version>
--        <exclusions>
--          <exclusion>
--            <groupId>javax.servlet</groupId>
--            <artifactId>jsp-api</artifactId>
--          </exclusion>
--          <exclusion>
--            <groupId>ant</groupId>
--            <artifactId>ant</artifactId>
--          </exclusion>
--        </exclusions>
 -      </dependency>
 -      <dependency>
 -        <groupId>tomcat</groupId>
--        <artifactId>jasper-runtime</artifactId>
+-        <artifactId>jasper-compiler</artifactId>
 -        <version>5.5.23</version>
-+        <groupId>org.glassfish.web</groupId>
-+        <artifactId>javax.servlet.jsp</artifactId>
-+        <version>2.2.5</version>
++        <groupId>org.apache.tomcat</groupId>
++        <artifactId>tomcat-jasper</artifactId>
++        <version>7.0.37</version>
+         <exclusions>
+           <exclusion>
+             <groupId>javax.servlet</groupId>
+@@ -465,17 +471,16 @@
+             <groupId>ant</groupId>
+             <artifactId>ant</artifactId>
+           </exclusion>
++          <exclusion>
++             <groupId>org.eclipse.jdt.core.compiler</groupId>
++             <artifactId>ecj</artifactId>
++          </exclusion>
+         </exclusions>
        </dependency>
        <dependency>
+-        <groupId>tomcat</groupId>
+-        <artifactId>jasper-runtime</artifactId>
+-        <version>5.5.23</version>
+-      </dependency>
+-      <dependency>
 -        <groupId>javax.servlet.jsp</groupId>
 -        <artifactId>jsp-api</artifactId>
 -        <version>2.1</version>
-+        <groupId>org.apache.tomcat</groupId>
-+        <artifactId>tomcat-servlet-api</artifactId>
-+        <version>7.0.37</version>
++        <groupId>org.eclipse.jetty</groupId>
++        <artifactId>test-jetty-servlet</artifactId>
++        <version>9.0.4.v20130625</version>
        </dependency>
        <dependency>
          <groupId>commons-el</groupId>
-@@ -483,6 +476,11 @@
+@@ -483,6 +488,11 @@
          <version>1.0</version>
        </dependency>
        <dependency>
@@ -2599,7 +2610,7 @@ index b7a7456..abdf900 100644
          <groupId>commons-logging</groupId>
          <artifactId>commons-logging</artifactId>
          <version>1.1.1</version>
-@@ -536,7 +534,7 @@
+@@ -536,7 +546,7 @@
        <dependency>
          <groupId>net.java.dev.jets3t</groupId>
          <artifactId>jets3t</artifactId>
@@ -2608,7 +2619,7 @@ index b7a7456..abdf900 100644
        </dependency>
        <dependency>
          <groupId>org.apache.mina</groupId>
-@@ -704,7 +702,7 @@
+@@ -704,7 +714,7 @@
        <dependency>
          <groupId>org.apache.bookkeeper</groupId>
          <artifactId>bookkeeper-server</artifactId>
@@ -2617,7 +2628,19 @@ index b7a7456..abdf900 100644
          <scope>compile</scope>
        </dependency>
        <dependency>
-@@ -797,9 +795,9 @@
+@@ -712,6 +722,11 @@
+         <artifactId>hsqldb</artifactId>
+         <version>2.0.0</version>
+       </dependency>
++      <dependency>
++        <groupId>com.google.code.findbugs</groupId>
++        <artifactId>jsr305</artifactId>
++        <version>1.3.9</version>
++      </dependency>
+     </dependencies>
+   </dependencyManagement>
+ 
+@@ -797,9 +812,9 @@
            <version>${avro.version}</version>
          </plugin>
          <plugin>
diff --git a/hadoop.spec b/hadoop.spec
index 6520ab6..6df9cfe 100644
--- a/hadoop.spec
+++ b/hadoop.spec
@@ -22,12 +22,12 @@
 %global httpfs_services hadoop-httpfs.service
 
 # Filter out undesired provides and requires
-%global __requires_exclude_from ^%{_libdir}/%{name}/libhadoop.so.*$
+%global __requires_exclude_from ^%{_libdir}/%{name}/libhadoop.so$
 %global __provides_exclude_from ^%{_libdir}/%{name}/.*$
 
 Name:   hadoop
 Version: 2.2.0
-Release: 1%{?dist}
+Release: 2%{?dist}
 Summary: A software platform for processing vast amounts of data
 # The BSD license file is missing
 # https://issues.apache.org/jira/browse/HADOOP-9849
@@ -50,13 +50,16 @@ Source12: hadoop-httpfs-env.sh
 Source13: hdfs-create-dirs
 # This patch includes the following upstream tickets:
 # https://issues.apache.org/jira/browse/HADOOP-9594
-# https://issues.apache.org/jira/browse/HADOOP-9605
-# https://issues.apache.org/jira/browse/HADOOP-9607
-# https://issues.apache.org/jira/browse/HADOOP-9610
+# https://issues.apache.org/jira/browse/MAPREDUCE-5431
 # https://issues.apache.org/jira/browse/HADOOP-9611
 # https://issues.apache.org/jira/browse/HADOOP-9613
 # https://issues.apache.org/jira/browse/HADOOP-9623
-# https://issues.apache.org/jira/browse/HADOOP-9650
+# https://issues.apache.org/jira/browse/HDFS-5411
+# https://issues.apache.org/jira/browse/HADOOP-10067
+# https://issues.apache.org/jira/browse/HDFS-5075
+# https://issues.apache.org/jira/browse/HADOOP-10068
+# https://issues.apache.org/jira/browse/HADOOP-10075
+# https://issues.apache.org/jira/browse/HADOOP-10076
 Patch0: hadoop-fedora-integration.patch
 # Fedora packaging guidelines for JNI library loading
 Patch2: hadoop-jni-library-loading.patch
@@ -134,7 +137,6 @@ BuildRequires: jettison
 BuildRequires: jetty
 BuildRequires: jetty-jspc-maven-plugin
 BuildRequires: jetty-util-ajax
-BuildRequires: jline
 BuildRequires: jsch
 BuildRequires: json_simple
 BuildRequires: jsr-305
@@ -191,13 +193,13 @@ BuildRequires: jersey-test-framework
 BuildRequires: maven-surefire-provider-junit4
 
 %description
-Hadoop is a framework that allows for the distributed processing of large data
-sets across clusters of computers using simple programming models.  It is
-designed to scale up from single servers to thousands of machines, each
+Apache Hadoop is a framework that allows for the distributed processing of
+large data sets across clusters of computers using simple programming models.
+It is designed to scale up from single servers to thousands of machines, each
 offering local computation and storage.
 
 %package client
-Summary: Libraries for Hadoop clients
+Summary: Libraries for Apache Hadoop clients
 Group: Applications/System
 BuildArch: noarch
 Requires: %{name}-common = %{version}-%{release}
@@ -206,15 +208,15 @@ Requires: %{name}-mapreduce = %{version}-%{release}
 Requires: %{name}-yarn = %{version}-%{release}
 
 %description client
-Hadoop is a framework that allows for the distributed processing of large data
-sets across clusters of computers using simple programming models.  It is
-designed to scale up from single servers to thousands of machines, each
+Apache Hadoop is a framework that allows for the distributed processing of
+large data sets across clusters of computers using simple programming models.
+It is designed to scale up from single servers to thousands of machines, each
 offering local computation and storage.
 
-This package provides libraries for Hadoop clients.
+This package provides libraries for Apache Hadoop clients.
 
 %package common
-Summary: Common files needed by Hadoop daemons
+Summary: Common files needed by Apache Hadoop daemons
 Group: Applications/System
 BuildArch: noarch
 Requires: /usr/sbin/useradd
@@ -255,38 +257,39 @@ Requires: txw2
 Requires: which
 
 %description common
-Hadoop is a framework that allows for the distributed processing of large data
-sets across clusters of computers using simple programming models.  It is
-designed to scale up from single servers to thousands of machines, each
+Apache Hadoop is a framework that allows for the distributed processing of
+large data sets across clusters of computers using simple programming models.
+It is designed to scale up from single servers to thousands of machines, each
 offering local computation and storage.
 
-This package contains common files and utilities needed by other Hadoop modules.
+This package contains common files and utilities needed by other Apache
+Hadoop modules.
 
 %package common-native
-Summary: The native Hadoop library file
+Summary: The native Apache Hadoop library file
 Group: Applications/System
 Requires: %{name}-common = %{version}-%{release}
 
 %description common-native
-Hadoop is a framework that allows for the distributed processing of large data
-sets across clusters of computers using simple programming models.  It is
-designed to scale up from single servers to thousands of machines, each
+Apache Hadoop is a framework that allows for the distributed processing of
+large data sets across clusters of computers using simple programming models.
+It is designed to scale up from single servers to thousands of machines, each
 offering local computation and storage.
 
 This package contains the native-hadoop library
 
 %if %{package_libhdfs}
 %package devel
-Summary: Headers for Hadoop
+Summary: Headers for Apache Hadoop
 Group: Development/System
 Requires: libhdfs%{?_isa} = %{version}-%{release}
 
 %description devel
-Header files for Hadoop's hdfs library and other utilities
+Header files for Apache Hadoop's hdfs library and other utilities
 %endif
 
 %package hdfs
-Summary: The Hadoop Distributed File System
+Summary: The Apache Hadoop Distributed File System
 Group: Applications/System
 BuildArch: noarch
 Requires: apache-commons-daemon-jsvc
@@ -297,17 +300,17 @@ Requires(preun): systemd
 Requires(postun): systemd
 
 %description hdfs
-Hadoop is a framework that allows for the distributed processing of large data
-sets across clusters of computers using simple programming models.  It is
-designed to scale up from single servers to thousands of machines, each
+Apache Hadoop is a framework that allows for the distributed processing of
+large data sets across clusters of computers using simple programming models.
+It is designed to scale up from single servers to thousands of machines, each
 offering local computation and storage.
 
 The Hadoop Distributed File System (HDFS) is the primary storage system
-used by Hadoop applications.
+used by Apache Hadoop applications.
 
 %if %{package_libhdfs}
 %package hdfs-fuse
-Summary: Allows mounting of Hadoop HDFS
+Summary: Allows mounting of Apache Hadoop HDFS
 Group: Development/Libraries
 Requires: fuse
 Requires: libhdfs%{?_isa} = %{version}-%{release}
@@ -317,9 +320,9 @@ Requires: %{name}-mapreduce = %{version}-%{release}
 Requires: %{name}-yarn = %{version}-%{release}
 
 %description hdfs-fuse
-Hadoop is a framework that allows for the distributed processing of large data
-sets across clusters of computers using simple programming models.  It is
-designed to scale up from single servers to thousands of machines, each
+Apache Hadoop is a framework that allows for the distributed processing of
+large data sets across clusters of computers using simple programming models.
+It is designed to scale up from single servers to thousands of machines, each
 offering local computation and storage.
 
 This package provides tools that allow HDFS to be mounted as a standard
@@ -341,9 +344,9 @@ Requires(preun): systemd
 Requires(postun): systemd
 
 %description httpfs
-Hadoop is a framework that allows for the distributed processing of large data
-sets across clusters of computers using simple programming models.  It is
-designed to scale up from single servers to thousands of machines, each
+Apache Hadoop is a framework that allows for the distributed processing of
+large data sets across clusters of computers using simple programming models.
+It is designed to scale up from single servers to thousands of machines, each
 offering local computation and storage.
 
 This package provides a server that provides HTTP REST API support for
@@ -354,32 +357,32 @@ the complete FileSystem/FileContext interface in HDFS.
 # most architectures so only generate on intel 64-bit
 %ifarch x86_64
 %package javadoc
-Summary: Javadoc for Hadoop
+Summary: Javadoc for Apache Hadoop
 Group: Documentation
 BuildArch: noarch
 
 %description javadoc
-This package contains the API documentation for %{name}
+This package contains the API documentation for %{name}.
 %endif
 
 %if %{package_libhdfs}
 %package -n libhdfs
-Summary: The Hadoop Filesystem Library
+Summary: The Apache Hadoop Filesystem Library
 Group: Development/Libraries
 Requires: %{name}-hdfs = %{version}-%{release}
 Requires: lzo
 
 %description -n libhdfs
-Hadoop is a framework that allows for the distributed processing of large data
-sets across clusters of computers using simple programming models.  It is
-designed to scale up from single servers to thousands of machines, each
+Apache Hadoop is a framework that allows for the distributed processing of
+large data sets across clusters of computers using simple programming models.
+It is designed to scale up from single servers to thousands of machines, each
 offering local computation and storage.
 
-This package provides the Hadoop Filesystem Library.
+This package provides the Apache Hadoop Filesystem Library.
 %endif
 
 %package mapreduce
-Summary: Hadoop MapReduce (MRv2)
+Summary: Apache Hadoop MapReduce (MRv2)
 Group: Applications/System
 BuildArch: noarch
 Requires(pre): %{name}-common = %{version}-%{release}
@@ -388,15 +391,15 @@ Requires(preun): systemd
 Requires(postun): systemd
 
 %description mapreduce
-Hadoop is a framework that allows for the distributed processing of large data
-sets across clusters of computers using simple programming models.  It is
-designed to scale up from single servers to thousands of machines, each
+Apache Hadoop is a framework that allows for the distributed processing of
+large data sets across clusters of computers using simple programming models.
+It is designed to scale up from single servers to thousands of machines, each
 offering local computation and storage.
 
-This package provides Hadoop MapReduce (MRv2).
+This package provides Apache Hadoop MapReduce (MRv2).
 
 %package mapreduce-examples
-Summary: Hadoop MapReduce (MRv2) examples
+Summary: Apache Hadoop MapReduce (MRv2) examples
 Group: Applications/System
 BuildArch: noarch
 Requires: %{name}-mapreduce = %{version}-%{release}
@@ -406,16 +409,16 @@ Requires: hsqldb
 This package contains mapreduce examples.
 
 %package maven-plugin
-Summary: Hadoop maven plugin
+Summary: Apache Hadoop maven plugin
 Group: Development/Libraries
 BuildArch: noarch
 Requires: maven
 
 %description maven-plugin
-The Hadoop maven plugin
+The Apache Hadoop maven plugin
 
 %package tests
-Summary: Hadoop test resources
+Summary: Apache Hadoop test resources
 BuildArch: noarch
 Requires: %{name}-common = %{version}-%{release}
 Requires: %{name}-hdfs = %{version}-%{release}
@@ -423,15 +426,15 @@ Requires: %{name}-mapreduce = %{version}-%{release}
 Requires: %{name}-yarn = %{version}-%{release}
 
 %description tests
-Hadoop is a framework that allows for the distributed processing of large data
-sets across clusters of computers using simple programming models.  It is
-designed to scale up from single servers to thousands of machines, each
+Apache Hadoop is a framework that allows for the distributed processing of
+large data sets across clusters of computers using simple programming models.
+It is designed to scale up from single servers to thousands of machines, each
 offering local computation and storage.
 
-This package contains test related resources for Hadoop.
+This package contains test related resources for Apache Hadoop.
 
 %package yarn
-Summary: Hadoop YARN
+Summary: Apache Hadoop YARN
 Group: Applications/System
 BuildArch: noarch
 Requires(pre): %{name}-common = %{version}-%{release}
@@ -443,26 +446,25 @@ Requires(preun): systemd
 Requires(postun): systemd
 
 %description yarn
-Hadoop is a framework that allows for the distributed processing of large data
-sets across clusters of computers using simple programming models.  It is
-designed to scale up from single servers to thousands of machines, each
+Apache Hadoop is a framework that allows for the distributed processing of
+large data sets across clusters of computers using simple programming models.
+It is designed to scale up from single servers to thousands of machines, each
 offering local computation and storage.
 
-This package contains Hadoop YARN.
+This package contains Apache Hadoop YARN.
 
 %package yarn-security
-Summary: The ability to run Hadoop YARN in secure mode
+Summary: The ability to run Apache Hadoop YARN in secure mode
 Group: Applications/System
 Requires: %{name}-yarn = %{version}-%{release}
 
 %description yarn-security
-Hadoop is a framework that allows for the distributed processing of large data
-sets across clusters of computers using simple programming models.  It is
-designed to scale up from single servers to thousands of machines, each
-offering local computation and storage.  YARN (Hadoop NextGen MapReduce) is
-a general purpose data-computation framework.
+Apache Hadoop is a framework that allows for the distributed processing of
+large data sets across clusters of computers using simple programming models.
+It is designed to scale up from single servers to thousands of machines, each
+offering local computation and storage.
 
-This package contains files needed to run Hadoop YARN in secure mode.
+This package contains files needed to run Apache Hadoop YARN in secure mode.
 
 %prep
 %setup -qn %{name}-common-%{commit}
@@ -525,33 +527,35 @@ This package contains files needed to run Hadoop YARN in secure mode.
 %mvn_package :%{name}-project-dist __noinstall
 
 # Create separate file lists for packaging
-%mvn_package :::tests: hadoop-tests
-%mvn_package :%{name}-client*::{}: hadoop-client
-%mvn_package :%{name}-hdfs*::{}: hadoop-hdfs
-%mvn_package :%{name}-mapreduce-examples*::{}: hadoop-mapreduce-examples
-%mvn_package :%{name}-mapreduce*::{}: hadoop-mapreduce
-%mvn_package :%{name}-archives::{}: hadoop-mapreduce
-%mvn_package :%{name}-datajoin::{}: hadoop-mapreduce
-%mvn_package :%{name}-distcp::{}: hadoop-mapreduce
-%mvn_package :%{name}-extras::{}: hadoop-mapreduce
-%mvn_package :%{name}-gridmix::{}: hadoop-mapreduce
-%mvn_package :%{name}-rumen::{}: hadoop-mapreduce
-%mvn_package :%{name}-streaming::{}: hadoop-mapreduce
-%mvn_package :%{name}-pipes::{}: hadoop-mapreduce
-%mvn_package :%{name}-tools*::{}: hadoop-mapreduce
-%mvn_package :%{name}-maven-plugins::{}: hadoop-maven-plugin
-%mvn_package :%{name}-minicluster::{}: hadoop-tests
-%mvn_package :%{name}-yarn*::{}: hadoop-yarn
+%mvn_package :::tests: %{name}-tests
+%mvn_package :%{name}-client*::{}: %{name}-client
+%mvn_package :%{name}-hdfs*::{}: %{name}-hdfs
+%mvn_package :%{name}-mapreduce-examples*::{}: %{name}-mapreduce-examples
+%mvn_package :%{name}-mapreduce*::{}: %{name}-mapreduce
+%mvn_package :%{name}-archives::{}: %{name}-mapreduce
+%mvn_package :%{name}-datajoin::{}: %{name}-mapreduce
+%mvn_package :%{name}-distcp::{}: %{name}-mapreduce
+%mvn_package :%{name}-extras::{}: %{name}-mapreduce
+%mvn_package :%{name}-gridmix::{}: %{name}-mapreduce
+%mvn_package :%{name}-rumen::{}: %{name}-mapreduce
+%mvn_package :%{name}-streaming::{}: %{name}-mapreduce
+%mvn_package :%{name}-pipes::{}: %{name}-mapreduce
+%mvn_package :%{name}-tools*::{}: %{name}-mapreduce
+%mvn_package :%{name}-maven-plugins::{}: %{name}-maven-plugin
+%mvn_package :%{name}-minicluster::{}: %{name}-tests
+%mvn_package :%{name}-yarn*::{}: %{name}-yarn
 
 # Jar files that need to be overridden due to installation location
-%mvn_file :%{name}-common::{}: %{_jnidir}/%{name}-common %{_datadir}/%{name}/common/%{name}-common
-%mvn_file :%{name}-common::tests: %{name}/%{name}-common-tests
+# Workaround for bz1023116
+#%%mvn_file :%{name}-common::{}: %{_jnidir}/%{name}-common %{_datadir}/%{name}/common/%{name}-common
+%mvn_file :%{name}-common::{}: %{_jnidir}/%{name}-common
+%mvn_file :%{name}-common::tests: %{name}/%{name}-common
 
 %build
 %ifnarch x86_64
 opts="-j"
 %endif
-%mvn_build $opts -- -Drequire.snappy=true -Dcontainer-executor.conf.dir=%{_sysconfdir}/hadoop -Pdist,native -DskipTests -DskipTest -DskipIT
+%mvn_build $opts -- -Drequire.snappy=true -Dcontainer-executor.conf.dir=%{_sysconfdir}/%{name} -Pdist,native -DskipTests -DskipTest -DskipIT
 
 # This takes a long time to run, so comment out for now
 #%%check
@@ -690,11 +694,11 @@ find %{buildroot} -name *.cmd | xargs rm -f
 
 # Modify hadoop-env.sh to point to correct locations for JAVA_HOME
 # and JSVC_HOME.
-sed -i "s|\${JAVA_HOME}|/usr/lib/jvm/jre|" %{buildroot}/%{_sysconfdir}/%{name}/hadoop-env.sh
-sed -i "s|\${JSVC_HOME}|/usr/bin|" %{buildroot}/%{_sysconfdir}/%{name}/hadoop-env.sh
+sed -i "s|\${JAVA_HOME}|/usr/lib/jvm/jre|" %{buildroot}/%{_sysconfdir}/%{name}/%{name}-env.sh
+sed -i "s|\${JSVC_HOME}|/usr/bin|" %{buildroot}/%{_sysconfdir}/%{name}/%{name}-env.sh
 
 # Ensure the java provided DocumentBuilderFactory is used
-sed -i "s|\(HADOOP_OPTS.*=.*\)\$HADOOP_CLIENT_OPTS|\1 -Djavax.xml.parsers.DocumentBuilderFactory=com.sun.org.apache.xerces.internal.jaxp.DocumentBuilderFactoryImpl \$HADOOP_CLIENT_OPTS|" %{buildroot}/%{_sysconfdir}/%{name}/hadoop-env.sh
+sed -i "s|\(HADOOP_OPTS.*=.*\)\$HADOOP_CLIENT_OPTS|\1 -Djavax.xml.parsers.DocumentBuilderFactory=com.sun.org.apache.xerces.internal.jaxp.DocumentBuilderFactoryImpl \$HADOOP_CLIENT_OPTS|" %{buildroot}/%{_sysconfdir}/%{name}/%{name}-env.sh
 echo "export YARN_OPTS=\"\$YARN_OPTS -Djavax.xml.parsers.DocumentBuilderFactory=com.sun.org.apache.xerces.internal.jaxp.DocumentBuilderFactoryImpl\"" >> %{buildroot}/%{_sysconfdir}/%{name}/yarn-env.sh
 
 # Workaround for bz1012059
@@ -702,6 +706,9 @@ install -pm 644 %{name}-project-dist/target/%{name}-project-dist-%{hadoop_versio
 install -pm 644 hadoop-project-dist/pom.xml %{buildroot}/%{_mavenpomdir}/JPP.%{name}-%{name}-project-dist.pom
 %add_maven_depmap JPP.%{name}-%{name}-project-dist.pom %{name}/%{name}-project-dist.jar
 
+# Workaround for bz1023116
+%{__ln_s} %{_jnidir}/%{name}-common.jar %{buildroot}/%{_datadir}/%{name}/common
+
 # client jar depenencies
 copy_dep_jars %{name}-client/target/%{name}-client-%{hadoop_version}/share/%{name}/client/lib %{buildroot}/%{_datadir}/%{name}/client/lib
 %{_bindir}/xmvn-subst %{buildroot}/%{_datadir}/%{name}/client/lib
@@ -746,7 +753,7 @@ cat > %{buildroot}/%{_sharedstatedir}/%{name}-httpfs/webapps/webhdfs/META-INF/co
 EOF
 
 # Remove the jars included in the webapp and create symlinks
-rm -f %{buildroot}%{_sharedstatedir}/%{name}-httpfs/webapps/webhdfs/WEB-INF/lib/hadoop-common*.jar
+rm -f %{buildroot}%{_sharedstatedir}/%{name}-httpfs/webapps/webhdfs/WEB-INF/lib/%{name}-common*.jar
 rm -f %{buildroot}%{_sharedstatedir}/%{name}-httpfs/webapps/webhdfs/WEB-INF/lib/tools*.jar
 %{_bindir}/xmvn-subst %{buildroot}/%{_sharedstatedir}/%{name}-httpfs/webapps/webhdfs/WEB-INF/lib
 %{__ln_s} %{_jnidir}/%{name}-common.jar %{buildroot}%{_sharedstatedir}/%{name}-httpfs/webapps/webhdfs/WEB-INF/lib
@@ -810,7 +817,7 @@ if [ "$lib" = "%_libdir" ]; then
   echo "_libdir is not located in /usr.  Lib location is wrong"
   exit 1
 fi
-sed -e "s|HADOOP_COMMON_LIB_NATIVE_DIR\s*=.*|HADOOP_COMMON_LIB_NATIVE_DIR=$lib/%{name}|" %{SOURCE1} > %{buildroot}/%{_libexecdir}/hadoop-layout.sh
+sed -e "s|HADOOP_COMMON_LIB_NATIVE_DIR\s*=.*|HADOOP_COMMON_LIB_NATIVE_DIR=$lib/%{name}|" %{SOURCE1} > %{buildroot}/%{_libexecdir}/%{name}-layout.sh
 
 # Default config
 cp -f %{SOURCE8} %{buildroot}/%{_sysconfdir}/%{name}/core-site.xml
@@ -853,11 +860,11 @@ do
 done
 
 # Ensure /var/run directories are recreated on boot
-echo "d %{_var}/run/%{name}-yarn 0775 yarn hadoop -" > %{buildroot}/%{_tmpfilesdir}/hadoop-yarn.conf
-echo "d %{_var}/run/%{name}-hdfs 0775 hdfs hadoop -" > %{buildroot}/%{_tmpfilesdir}/hadoop-hdfs.conf
-echo "d %{_var}/run/%{name}-mapreduce 0775 mapred hadoop -" > %{buildroot}/%{_tmpfilesdir}/hadoop-mapreduce.conf
+echo "d %{_var}/run/%{name}-yarn 0775 yarn hadoop -" > %{buildroot}/%{_tmpfilesdir}/%{name}-yarn.conf
+echo "d %{_var}/run/%{name}-hdfs 0775 hdfs hadoop -" > %{buildroot}/%{_tmpfilesdir}/%{name}-hdfs.conf
+echo "d %{_var}/run/%{name}-mapreduce 0775 mapred hadoop -" > %{buildroot}/%{_tmpfilesdir}/%{name}-mapreduce.conf
 %if %{package_httpfs}
-echo "d %{_var}/run/%{name}-httpfs 0775 httpfs hadoop -" > %{buildroot}/%{_tmpfilesdir}/hadoop-httpfs.conf
+echo "d %{_var}/run/%{name}-httpfs 0775 httpfs hadoop -" > %{buildroot}/%{_tmpfilesdir}/%{name}-httpfs.conf
 %endif
 
 # logrotate config
@@ -889,21 +896,21 @@ getent group hadoop >/dev/null || groupadd -r hadoop
 
 %pre hdfs
 getent group hdfs >/dev/null || groupadd -r hdfs
-getent passwd hdfs >/dev/null || /usr/sbin/useradd --comment "Hadoop HDFS" --shell /sbin/nologin -M -r -g hdfs -G hadoop --home %{_var}/cache/%{name}-hdfs hdfs
+getent passwd hdfs >/dev/null || /usr/sbin/useradd --comment "Apache Hadoop HDFS" --shell /sbin/nologin -M -r -g hdfs -G hadoop --home %{_var}/cache/%{name}-hdfs hdfs
 
 %if %{package_httpfs}
 %pre httpfs 
 getent group httpfs >/dev/null || groupadd -r httpfs
-getent passwd httpfs >/dev/null || /usr/sbin/useradd --comment "Hadoop HTTPFS" --shell /sbin/nologin -M -r -g httpfs -G httpfs --home %{_var}/run/%{name}-httpfs httpfs
+getent passwd httpfs >/dev/null || /usr/sbin/useradd --comment "Apache Hadoop HTTPFS" --shell /sbin/nologin -M -r -g httpfs -G httpfs --home %{_var}/run/%{name}-httpfs httpfs
 %endif
 
 %pre mapreduce
 getent group mapred >/dev/null || groupadd -r mapred
-getent passwd mapred >/dev/null || /usr/sbin/useradd --comment "Hadoop MapReduce" --shell /sbin/nologin -M -r -g mapred -G hadoop --home %{_var}/cache/%{name}-mapreduce mapred
+getent passwd mapred >/dev/null || /usr/sbin/useradd --comment "Apache Hadoop MapReduce" --shell /sbin/nologin -M -r -g mapred -G hadoop --home %{_var}/cache/%{name}-mapreduce mapred
 
 %pre yarn
 getent group yarn >/dev/null || groupadd -r yarn
-getent passwd yarn >/dev/null || /usr/sbin/useradd --comment "Hadoop Yarn" --shell /sbin/nologin -M -r -g yarn -G hadoop --home %{_var}/cache/%{name}-yarn yarn
+getent passwd yarn >/dev/null || /usr/sbin/useradd --comment "Apache Hadoop Yarn" --shell /sbin/nologin -M -r -g yarn -G hadoop --home %{_var}/cache/%{name}-yarn yarn
 
 %preun hdfs
 %systemd_preun %{hdfs_services}
@@ -919,7 +926,7 @@ getent passwd yarn >/dev/null || /usr/sbin/useradd --comment "Hadoop Yarn" --she
 %preun yarn
 %systemd_preun %{yarn_services}
 
-%post common -p /sbin/ldconfig
+%post common-native -p /sbin/ldconfig
 
 %post hdfs
 %systemd_post %{hdfs_services}
@@ -939,7 +946,7 @@ getent passwd yarn >/dev/null || /usr/sbin/useradd --comment "Hadoop Yarn" --she
 %post yarn
 %systemd_post %{yarn_services}
 
-%postun common -p /sbin/ldconfig
+%postun common-native -p /sbin/ldconfig
 
 %postun hdfs
 %systemd_postun_with_restart %{hdfs_services}
@@ -959,7 +966,7 @@ getent passwd yarn >/dev/null || /usr/sbin/useradd --comment "Hadoop Yarn" --she
 %postun yarn
 %systemd_postun_with_restart %{yarn_services}
 
-%files -f .mfiles-hadoop-client client
+%files -f .mfiles-%{name}-client client
 %{_datadir}/%{name}/client
 
 %files -f .mfiles common
@@ -976,6 +983,10 @@ getent passwd yarn >/dev/null || /usr/sbin/useradd --comment "Hadoop Yarn" --she
 %config(noreplace) %{_sysconfdir}/%{name}/ssl-server.xml.example
 %dir %{_datadir}/%{name}
 %dir %{_datadir}/%{name}/common
+
+# Workaround for bz1023116
+%{_datadir}/%{name}/common/%{name}-common.jar
+
 %{_datadir}/%{name}/common/lib
 %{_libexecdir}/%{name}-config.sh
 %{_libexecdir}/%{name}-layout.sh
@@ -1001,7 +1012,7 @@ getent passwd yarn >/dev/null || /usr/sbin/useradd --comment "Hadoop Yarn" --she
 %{_libdir}/libhdfs.so
 %endif
 
-%files -f .mfiles-hadoop-hdfs hdfs
+%files -f .mfiles-%{name}-hdfs hdfs
 %exclude %{_datadir}/%{name}/client
 %config(noreplace) %{_sysconfdir}/%{name}/hdfs-site.xml
 %config(noreplace) %{_sysconfdir}/security/limits.d/hdfs.conf
@@ -1066,7 +1077,7 @@ getent passwd yarn >/dev/null || /usr/sbin/useradd --comment "Hadoop Yarn" --she
 %{_libdir}/libhdfs.so.*
 %endif
 
-%files -f .mfiles-hadoop-mapreduce mapreduce
+%files -f .mfiles-%{name}-mapreduce mapreduce
 %exclude %{_datadir}/%{name}/client
 %config(noreplace) %{_sysconfdir}/%{name}/mapred-env.sh
 %config(noreplace) %{_sysconfdir}/%{name}/mapred-queues.xml.template
@@ -1084,14 +1095,14 @@ getent passwd yarn >/dev/null || /usr/sbin/useradd --comment "Hadoop Yarn" --she
 %attr(0755,mapred,hadoop) %dir %{_var}/log/%{name}-mapreduce
 %attr(0755,mapred,hadoop) %dir %{_var}/cache/%{name}-mapreduce
 
-%files -f .mfiles-hadoop-mapreduce-examples mapreduce-examples
+%files -f .mfiles-%{name}-mapreduce-examples mapreduce-examples
 
-%files -f .mfiles-hadoop-maven-plugin maven-plugin
+%files -f .mfiles-%{name}-maven-plugin maven-plugin
 %doc hadoop-dist/target/hadoop-%{hadoop_version}/share/doc/hadoop/common/LICENSE.txt
 
-%files -f .mfiles-hadoop-tests tests
+%files -f .mfiles-%{name}-tests tests
 
-%files -f .mfiles-hadoop-yarn yarn
+%files -f .mfiles-%{name}-yarn yarn
 %exclude %{_datadir}/%{name}/client
 %config(noreplace) %{_sysconfdir}/%{name}/capacity-scheduler.xml
 %config(noreplace) %{_sysconfdir}/%{name}/yarn-env.sh
@@ -1119,6 +1130,13 @@ getent passwd yarn >/dev/null || /usr/sbin/useradd --comment "Hadoop Yarn" --she
 %attr(6050,root,yarn) %{_bindir}/container-executor
 
 %changelog
+* Tue Dec  3 2013 Robert Rati <rrati at redhat> - 2.2.0-2
+- Changed provides filter to just filter the .so
+- Corrected naming of hadoop-common test jar
+- Removed jline BuildRequires
+- Moved pre/port install invocation of ldconfig to common-native
+- Added workaround for bz1023116
+
 * Wed Oct 23 2013 Robert Rati <rrati at redhat> - 2.2.0-1
 - Update to upstream 2.2.0
 - New patch to open libjvm with dlopen


More information about the scm-commits mailing list