[hadoop] Build 2.0.5-8
rrati
rrati at fedoraproject.org
Wed Aug 28 13:27:51 UTC 2013
commit 4c8acc0e49d4030f7802fe1b102fa5db3f69bfa8
Author: Robert Rati <rrati at redhat.com>
Date: Wed Aug 28 09:28:23 2013 -0400
Build 2.0.5-8
hadoop-fedora-integration.patch | 97 ++++++++++++++++-----------------------
hadoop.spec | 5 ++-
2 files changed, 44 insertions(+), 58 deletions(-)
---
diff --git a/hadoop-fedora-integration.patch b/hadoop-fedora-integration.patch
index eedd056..46a12b1 100644
--- a/hadoop-fedora-integration.patch
+++ b/hadoop-fedora-integration.patch
@@ -50,20 +50,23 @@ index 36a93a7..c2b52c9 100644
<exclusion>
<groupId>com.sun.jersey</groupId>
diff --git a/hadoop-common-project/hadoop-annotations/pom.xml b/hadoop-common-project/hadoop-annotations/pom.xml
-index 6996ea3..144f7d9 100644
+index 6996ea3..6750191 100644
--- a/hadoop-common-project/hadoop-annotations/pom.xml
+++ b/hadoop-common-project/hadoop-annotations/pom.xml
-@@ -48,8 +48,8 @@
+@@ -48,11 +48,8 @@
</activation>
<dependencies>
<dependency>
- <groupId>jdk.tools</groupId>
- <artifactId>jdk.tools</artifactId>
+- <version>1.6</version>
+- <scope>system</scope>
+- <systemPath>${java.home}/../lib/tools.jar</systemPath>
+ <groupId>com.sun</groupId>
+ <artifactId>tools</artifactId>
- <version>1.6</version>
- <scope>system</scope>
- <systemPath>${java.home}/../lib/tools.jar</systemPath>
+ </dependency>
+ </dependencies>
+ </profile>
diff --git a/hadoop-common-project/hadoop-auth/pom.xml b/hadoop-common-project/hadoop-auth/pom.xml
index a8642c0..29b4898 100644
--- a/hadoop-common-project/hadoop-auth/pom.xml
@@ -419,10 +422,10 @@ index 9e318ae..949db05 100644
/**
* General servlet which is admin-authorized.
diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpServer.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpServer.java
-index 7b5b17e..2e7bf53 100644
+index 7b5b17e..009403b 100644
--- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpServer.java
+++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpServer.java
-@@ -60,27 +60,31 @@
+@@ -60,27 +60,28 @@
import org.apache.hadoop.security.authorize.AccessControlList;
import org.apache.hadoop.security.ssl.SSLFactory;
import org.apache.hadoop.util.ReflectionUtils;
@@ -444,7 +447,6 @@ index 7b5b17e..2e7bf53 100644
-import org.mortbay.jetty.webapp.WebAppContext;
-import org.mortbay.thread.QueuedThreadPool;
-import org.mortbay.util.MultiException;
-+//import org.eclipse.jetty.io.Buffer;
+import org.eclipse.jetty.server.Handler;
+import org.eclipse.jetty.http.MimeTypes;
+import org.eclipse.jetty.server.HttpConfiguration;
@@ -453,8 +455,6 @@ index 7b5b17e..2e7bf53 100644
+import org.eclipse.jetty.server.ServerConnector;
+import org.eclipse.jetty.server.handler.ContextHandler;
+import org.eclipse.jetty.server.handler.ContextHandlerCollection;
-+//import org.eclipse.jetty.server.nio.SelectChannelConnector;
-+//import org.eclipse.jetty.server.ssl.SslSocketConnector;
+import org.eclipse.jetty.servlet.ServletContextHandler;
+import org.eclipse.jetty.servlet.DefaultServlet;
+import org.eclipse.jetty.servlet.FilterHolder;
@@ -472,7 +472,7 @@ index 7b5b17e..2e7bf53 100644
/**
* Create a Jetty embedded server to answer http requests. The primary goal
* is to serve up status information for the server.
-@@ -111,11 +115,12 @@
+@@ -111,11 +112,12 @@
private SSLFactory sslFactory;
protected final Server webServer;
@@ -488,7 +488,7 @@ index 7b5b17e..2e7bf53 100644
protected final List<String> filterNames = new ArrayList<String>();
private static final int MAX_RETRIES = 10;
static final String STATE_DESCRIPTION_ALIVE = " - alive";
-@@ -126,12 +131,12 @@
+@@ -126,12 +128,12 @@
/** Same as this(name, bindAddress, port, findPort, null); */
public HttpServer(String name, String bindAddress, int port, boolean findPort
) throws IOException {
@@ -504,7 +504,7 @@ index 7b5b17e..2e7bf53 100644
}
/**
-@@ -150,7 +155,7 @@ public HttpServer(String name, String bindAddress, int port,
+@@ -150,7 +152,7 @@ public HttpServer(String name, String bindAddress, int port,
*/
public HttpServer(String name, String bindAddress, int port,
boolean findPort, Configuration conf, String[] pathSpecs) throws IOException {
@@ -513,7 +513,7 @@ index 7b5b17e..2e7bf53 100644
}
/**
-@@ -164,13 +169,13 @@ public HttpServer(String name, String bindAddress, int port,
+@@ -164,13 +166,13 @@ public HttpServer(String name, String bindAddress, int port,
*/
public HttpServer(String name, String bindAddress, int port,
boolean findPort, Configuration conf) throws IOException {
@@ -529,7 +529,7 @@ index 7b5b17e..2e7bf53 100644
}
/**
-@@ -186,8 +191,8 @@ public HttpServer(String name, String bindAddress, int port,
+@@ -186,8 +188,8 @@ public HttpServer(String name, String bindAddress, int port,
*/
public HttpServer(String name, String bindAddress, int port,
boolean findPort, Configuration conf, AccessControlList adminsAcl,
@@ -540,7 +540,7 @@ index 7b5b17e..2e7bf53 100644
}
/**
-@@ -206,11 +211,25 @@ public HttpServer(String name, String bindAddress, int port,
+@@ -206,11 +208,17 @@ public HttpServer(String name, String bindAddress, int port,
*/
public HttpServer(String name, String bindAddress, int port,
boolean findPort, Configuration conf, AccessControlList adminsAcl,
@@ -548,14 +548,6 @@ index 7b5b17e..2e7bf53 100644
- webServer = new Server();
+ ServerConnector connector, String[] pathSpecs,
+ Server server) throws IOException {
-+// int maxThreads = conf.getInt(HTTP_MAX_THREADS, -1);
-+// // If HTTP_MAX_THREADS is not configured, QueueThreadPool() will use the
-+// // default value (currently 250).
-+// QueuedThreadPool threadPool = maxThreads == -1 ?
-+// new QueuedThreadPool() : new QueuedThreadPool(maxThreads);
-+// threadPool.setDaemon(true);
-+//
-+// webServer = new Server(threadPool);
this.findPort = findPort;
this.adminsAcl = adminsAcl;
@@ -568,7 +560,7 @@ index 7b5b17e..2e7bf53 100644
if(connector == null) {
listenerStartedExternally = false;
if (HttpConfig.isSecure()) {
-@@ -220,11 +239,18 @@ public HttpServer(String name, String bindAddress, int port,
+@@ -220,11 +228,18 @@ public HttpServer(String name, String bindAddress, int port,
} catch (GeneralSecurityException ex) {
throw new IOException(ex);
}
@@ -592,7 +584,7 @@ index 7b5b17e..2e7bf53 100644
};
listener = sslListener;
} else {
-@@ -239,17 +265,8 @@ protected SSLServerSocketFactory createFactory() throws Exception {
+@@ -239,17 +254,8 @@ protected SSLServerSocketFactory createFactory() throws Exception {
webServer.addConnector(listener);
@@ -611,7 +603,7 @@ index 7b5b17e..2e7bf53 100644
webAppContext = new WebAppContext();
webAppContext.setDisplayName(name);
-@@ -258,7 +275,8 @@ protected SSLServerSocketFactory createFactory() throws Exception {
+@@ -258,7 +264,8 @@ protected SSLServerSocketFactory createFactory() throws Exception {
webAppContext.getServletContext().setAttribute(CONF_CONTEXT_ATTRIBUTE, conf);
webAppContext.getServletContext().setAttribute(ADMINS_ACL, adminsAcl);
addNoCacheFilter(webAppContext);
@@ -621,7 +613,7 @@ index 7b5b17e..2e7bf53 100644
addDefaultApps(contexts, appDir, conf);
-@@ -293,19 +311,38 @@ private void addNoCacheFilter(WebAppContext ctxt) {
+@@ -293,19 +300,30 @@ private void addNoCacheFilter(WebAppContext ctxt) {
* provided. This wrapper and all subclasses must create at least one
* listener.
*/
@@ -648,14 +640,6 @@ index 7b5b17e..2e7bf53 100644
+ conn.setAcceptQueueSize(128);
+ conn.setIdleTimeout(10000);
+ return conn;
-+
-+// SelectChannelConnector ret = new SelectChannelConnector();
-+// ret.setLowResourceMaxIdleTime(10000);
-+// ret.setAcceptQueueSize(128);
-+// ret.setResolveNames(false);
-+// ret.setUseDirectBuffers(false);
-+// ret.setRequestHeaderSize(1024*64);
-+// return ret;
+ }
+
+ @InterfaceAudience.Private
@@ -670,7 +654,7 @@ index 7b5b17e..2e7bf53 100644
}
/** Get an array of FilterConfiguration specified in the conf */
-@@ -337,14 +374,14 @@ protected void addDefaultApps(ContextHandlerCollection parent,
+@@ -337,14 +355,14 @@ protected void addDefaultApps(ContextHandlerCollection parent,
// set up the context for "/logs/" if "hadoop.log.dir" property is defined.
String logDir = System.getProperty("hadoop.log.dir");
if (logDir != null) {
@@ -687,7 +671,7 @@ index 7b5b17e..2e7bf53 100644
}
logContext.setDisplayName("logs");
setContextAttributes(logContext, conf);
-@@ -352,7 +389,7 @@ protected void addDefaultApps(ContextHandlerCollection parent,
+@@ -352,7 +370,7 @@ protected void addDefaultApps(ContextHandlerCollection parent,
defaultContexts.put(logContext, true);
}
// set up the context for "/static/*"
@@ -696,7 +680,7 @@ index 7b5b17e..2e7bf53 100644
staticContext.setResourceBase(appDir + "/static");
staticContext.addServlet(DefaultServlet.class, "/*");
staticContext.setDisplayName("static");
-@@ -360,7 +397,7 @@ protected void addDefaultApps(ContextHandlerCollection parent,
+@@ -360,7 +378,7 @@ protected void addDefaultApps(ContextHandlerCollection parent,
defaultContexts.put(staticContext, true);
}
@@ -705,7 +689,7 @@ index 7b5b17e..2e7bf53 100644
context.getServletContext().setAttribute(CONF_CONTEXT_ATTRIBUTE, conf);
context.getServletContext().setAttribute(ADMINS_ACL, adminsAcl);
}
-@@ -377,10 +414,11 @@ protected void addDefaultServlets() {
+@@ -377,10 +395,11 @@ protected void addDefaultServlets() {
addServlet("conf", "/conf", ConfServlet.class);
}
@@ -719,7 +703,7 @@ index 7b5b17e..2e7bf53 100644
defaultContexts.put(ctxt, isFiltered);
}
-@@ -481,7 +519,7 @@ public void addInternalServlet(String name, String pathSpec,
+@@ -481,7 +500,7 @@ public void addInternalServlet(String name, String pathSpec,
FilterMapping fmap = new FilterMapping();
fmap.setPathSpec(pathSpec);
fmap.setFilterName(SPNEGO_FILTER);
@@ -728,7 +712,7 @@ index 7b5b17e..2e7bf53 100644
handler.addFilterMapping(fmap);
}
}
-@@ -495,9 +533,9 @@ public void addFilter(String name, String classname,
+@@ -495,9 +514,9 @@ public void addFilter(String name, String classname,
LOG.info("Added filter " + name + " (class=" + classname
+ ") to context " + webAppContext.getDisplayName());
final String[] ALL_URLS = { "/*" };
@@ -740,7 +724,7 @@ index 7b5b17e..2e7bf53 100644
defineFilter(ctx, name, classname, parameters, ALL_URLS);
LOG.info("Added filter " + name + " (class=" + classname
+ ") to context " + ctx.getDisplayName());
-@@ -511,7 +549,7 @@ public void addGlobalFilter(String name, String classname,
+@@ -511,7 +530,7 @@ public void addGlobalFilter(String name, String classname,
Map<String, String> parameters) {
final String[] ALL_URLS = { "/*" };
defineFilter(webAppContext, name, classname, parameters, ALL_URLS);
@@ -749,7 +733,7 @@ index 7b5b17e..2e7bf53 100644
defineFilter(ctx, name, classname, parameters, ALL_URLS);
}
LOG.info("Added global filter '" + name + "' (class=" + classname + ")");
-@@ -520,16 +558,18 @@ public void addGlobalFilter(String name, String classname,
+@@ -520,16 +539,18 @@ public void addGlobalFilter(String name, String classname,
/**
* Define a filter for a context and set up default url mappings.
*/
@@ -771,7 +755,7 @@ index 7b5b17e..2e7bf53 100644
fmap.setFilterName(name);
ServletHandler handler = ctx.getServletHandler();
handler.addFilter(holder, fmap);
-@@ -541,13 +581,13 @@ protected void defineFilter(Context ctx, String name,
+@@ -541,13 +562,13 @@ protected void defineFilter(Context ctx, String name,
* @param webAppCtx The WebApplicationContext to add to
*/
protected void addFilterPathMapping(String pathSpec,
@@ -787,7 +771,7 @@ index 7b5b17e..2e7bf53 100644
handler.addFilterMapping(fmap);
}
}
-@@ -581,7 +621,7 @@ protected String getWebAppsPath(String appName) throws FileNotFoundException {
+@@ -581,7 +602,7 @@ protected String getWebAppsPath(String appName) throws FileNotFoundException {
* @return the port
*/
public int getPort() {
@@ -796,7 +780,7 @@ index 7b5b17e..2e7bf53 100644
}
/**
-@@ -607,12 +647,12 @@ public void addSslListener(InetSocketAddress addr, String keystore,
+@@ -607,12 +628,12 @@ public void addSslListener(InetSocketAddress addr, String keystore,
if (webServer.isStarted()) {
throw new IOException("Failed to add ssl listener");
}
@@ -813,7 +797,7 @@ index 7b5b17e..2e7bf53 100644
webServer.addConnector(sslListener);
}
-@@ -636,14 +676,14 @@ public void addSslListener(InetSocketAddress addr, Configuration sslConf,
+@@ -636,14 +657,14 @@ public void addSslListener(InetSocketAddress addr, Configuration sslConf,
System.setProperty("javax.net.ssl.trustStoreType", sslConf.get(
"ssl.server.truststore.type", "jks"));
}
@@ -834,7 +818,7 @@ index 7b5b17e..2e7bf53 100644
webServer.addConnector(sslListener);
}
-@@ -1075,10 +1115,10 @@ public void doFilter(ServletRequest request,
+@@ -1075,10 +1096,10 @@ public void doFilter(ServletRequest request,
*/
private String inferMimeType(ServletRequest request) {
String path = ((HttpServletRequest)request).getRequestURI();
@@ -1855,10 +1839,10 @@ index 0be46de..b2834b8 100644
if (conf.getBoolean(DFS_HTTPS_ENABLE_KEY, false)) {
boolean needClientAuth = conf.getBoolean(DFS_CLIENT_HTTPS_NEED_AUTH_KEY,
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/SecureDataNodeStarter.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/SecureDataNodeStarter.java
-index 0fda306..a1f6d36 100644
+index 0fda306..77c6c82 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/SecureDataNodeStarter.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/SecureDataNodeStarter.java
-@@ -32,14 +32,16 @@
+@@ -32,14 +32,15 @@
import org.apache.hadoop.http.HttpServer;
import org.apache.hadoop.security.UserGroupInformation;
import org.apache.hadoop.security.ssl.SSLFactory;
@@ -1867,7 +1851,6 @@ index 0fda306..a1f6d36 100644
-import org.mortbay.jetty.security.SslSocketConnector;
+import org.eclipse.jetty.server.Server;
+import org.eclipse.jetty.server.ServerConnector;
-+//import org.eclipse.jetty.server.ssl.SslSocketConnector;
+import org.eclipse.jetty.util.ssl.SslContextFactory;
import javax.net.ssl.SSLServerSocketFactory;
@@ -1878,7 +1861,7 @@ index 0fda306..a1f6d36 100644
/**
* Utility class to start a datanode in a secure cluster, first obtaining
* privileged resources before main startup and handing them to the datanode.
-@@ -50,17 +52,21 @@
+@@ -50,17 +51,21 @@
*/
public static class SecureResources {
private final ServerSocket streamingSocket;
@@ -1903,7 +1886,7 @@ index 0fda306..a1f6d36 100644
}
private String [] args;
-@@ -94,6 +100,9 @@ public void start() throws Exception {
+@@ -94,6 +99,9 @@ public void start() throws Exception {
@VisibleForTesting
public static SecureResources getSecureResources(final SSLFactory sslFactory,
Configuration conf) throws Exception {
@@ -1913,7 +1896,7 @@ index 0fda306..a1f6d36 100644
// Obtain secure port for data streaming to datanode
InetSocketAddress streamingAddr = DataNode.getStreamingAddr(conf);
int socketWriteTimeout = conf.getInt(DFSConfigKeys.DFS_DATANODE_SOCKET_WRITE_TIMEOUT_KEY,
-@@ -110,22 +119,28 @@ public static SecureResources getSecureResources(final SSLFactory sslFactory,
+@@ -110,22 +118,28 @@ public static SecureResources getSecureResources(final SSLFactory sslFactory,
}
// Obtain secure listener for web server
@@ -1949,7 +1932,7 @@ index 0fda306..a1f6d36 100644
}
InetSocketAddress infoSocAddr = DataNode.getInfoAddr(conf);
-@@ -138,7 +153,7 @@ protected SSLServerSocketFactory createFactory() throws Exception {
+@@ -138,7 +152,7 @@ protected SSLServerSocketFactory createFactory() throws Exception {
"context. Needed " + streamingAddr.getPort() + ", got " + ss.getLocalPort());
}
System.err.println("Successfully obtained privileged resources (streaming port = "
@@ -1958,7 +1941,7 @@ index 0fda306..a1f6d36 100644
if ((ss.getLocalPort() > 1023 || listener.getPort() > 1023) &&
UserGroupInformation.isSecurityEnabled()) {
-@@ -146,7 +161,7 @@ protected SSLServerSocketFactory createFactory() throws Exception {
+@@ -146,7 +160,7 @@ protected SSLServerSocketFactory createFactory() throws Exception {
}
System.err.println("Opened streaming server at " + streamingAddr);
System.err.println("Opened info server at " + infoSocAddr);
diff --git a/hadoop.spec b/hadoop.spec
index ab92f4c..405d8c2 100644
--- a/hadoop.spec
+++ b/hadoop.spec
@@ -16,7 +16,7 @@
Name: hadoop
Version: 2.0.5
-Release: 7%{?dist}
+Release: 8%{?dist}
Summary: A software platform for processing vast amounts of data
# The BSD license file is missing
# https://issues.apache.org/jira/browse/HADOOP-9849
@@ -963,6 +963,9 @@ getent passwd yarn >/dev/null || /usr/sbin/useradd --comment "Hadoop Yarn" --she
%attr(6050,root,yarn) %{_bindir}/container-executor
%changelog
+* Wed Aug 28 2013 Robert Rati <rrati at redhat> - 2.0.5-8
+- Removed systemPath, version, and scope from tools.jar dependency definition
+
* Tue Aug 20 2013 Robert Rati <rrati at redhat> - 2.0.5-7
- Changed hdfs subpackage from hadoop-libhdfs to libhdfs
- Don't build any packages on arm architectures
More information about the scm-commits
mailing list