Browse Source

[hadoop] Set MR2 as default

Keep backward compatibility by not switching current installations to MR2.
Romain Rigaux 12 years ago
parent
commit
59a02b2

+ 44 - 40
desktop/conf.dist/hue.ini

@@ -277,7 +277,7 @@
 
     # Have Hue initiated authn requests be signed and provide a certificate.
     ## authn_requests_signed=false
-    
+
     # Have Hue initiated logout requests be signed and provide a certificate.
     ## logout_requests_signed=false
 
@@ -316,28 +316,28 @@
       # Defaults to $HADOOP_CONF_DIR or /etc/hadoop/conf
       ## hadoop_conf_dir=/etc/hadoop/conf
 
-  # Configuration for MapReduce 0.20 JobTracker (MR1)
+  # Configuration for YARN (MR2)
   # ------------------------------------------------------------------------
-  [[mapred_clusters]]
+  [[yarn_clusters]]
 
     [[[default]]]
-      # Enter the host on which you are running the Hadoop JobTracker
-      jobtracker_host=localhost
-      # The port where the JobTracker IPC listens on
-      jobtracker_port=8021
-      # Thrift plug-in port for the JobTracker
-      ## thrift_port=9290
+      # Enter the host on which you are running the ResourceManager
+      ## resourcemanager_host=localhost
+
+      # The port where the ResourceManager IPC listens on
+      ## resourcemanager_port=8032
+
       # Whether to submit jobs to this cluster
-      ## submit_to=True
+      submit_to=True
 
-      # Change this if your MapReduce cluster is Kerberos-secured
+      # Change this if your YARN cluster is Kerberos-secured
       ## security_enabled=false
 
-      # Settings about this MR1 cluster. If you install MR1 in a
+      # Settings about this MR2 cluster. If you install MR2 in a
       # different location, you need to set the following.
 
-      # Defaults to $HADOOP_MR1_HOME or /usr/lib/hadoop-0.20-mapreduce
-      ## hadoop_mapred_home=/usr/lib/hadoop-0.20-mapreduce
+      # Defaults to $HADOOP_MR2_HOME or /usr/lib/hadoop-mapreduce
+      ## hadoop_mapred_home=/usr/lib/hadoop-mapreduce
 
       # Defaults to $HADOOP_BIN or /usr/bin/hadoop
       ## hadoop_bin=/usr/bin/hadoop
@@ -345,34 +345,43 @@
       # Defaults to $HADOOP_CONF_DIR or /etc/hadoop/conf
       ## hadoop_conf_dir=/etc/hadoop/conf
 
-    # HA support by specifying multiple clusters
-    # e.g.
+      # URL of the ResourceManager API
+      ## resourcemanager_api_url=http://localhost:8088
 
-    # [[[ha]]]
-      # Enter the host on which you are running the failover JobTracker
-      # jobtracker_host=localhost-ha
+      # URL of the ProxyServer API
+      ## proxy_api_url=http://localhost:8088
 
+      # URL of the HistoryServer API
+      # history_server_api_url=http://localhost:19888
 
-  # Configuration for YARN (MR2)
+      # URL of the NodeManager API
+      # node_manager_api_url=http://localhost:8042
+
+  # Configuration for MapReduce (MR1)
   # ------------------------------------------------------------------------
-  [[yarn_clusters]]
+  [[mapred_clusters]]
 
     [[[default]]]
-      # Enter the host on which you are running the ResourceManager
-      resourcemanager_host=localhost
-      # The port where the ResourceManager IPC listens on
-      resourcemanager_port=8032
+      # Enter the host on which you are running the Hadoop JobTracker
+      ## jobtracker_host=localhost
+
+      # The port where the JobTracker IPC listens on
+      ## jobtracker_port=8021
+
+      # Thrift plug-in port for the JobTracker
+      ## thrift_port=9290
+
       # Whether to submit jobs to this cluster
-      ## submit_to=False
+      submit_to=False
 
-      # Change this if your YARN cluster is Kerberos-secured
+      # Change this if your MapReduce cluster is Kerberos-secured
       ## security_enabled=false
 
-      # Settings about this MR2 cluster. If you install MR2 in a
+      # Settings about this MR1 cluster. If you install MR1 in a
       # different location, you need to set the following.
 
-      # Defaults to $HADOOP_MR2_HOME or /usr/lib/hadoop-mapreduce
-      ## hadoop_mapred_home=/usr/lib/hadoop-mapreduce
+      # Defaults to $HADOOP_MR1_HOME or /usr/lib/hadoop-0.20-mapreduce
+      ## hadoop_mapred_home=/usr/lib/hadoop-0.20-mapreduce
 
       # Defaults to $HADOOP_BIN or /usr/bin/hadoop
       ## hadoop_bin=/usr/bin/hadoop
@@ -380,17 +389,12 @@
       # Defaults to $HADOOP_CONF_DIR or /etc/hadoop/conf
       ## hadoop_conf_dir=/etc/hadoop/conf
 
-      # URL of the ResourceManager API
-      ## resourcemanager_api_url=http://localhost:8088
-
-      # URL of the ProxyServer API
-      ## proxy_api_url=http://localhost:8088
-
-      # URL of the HistoryServer API
-      history_server_api_url=http://localhost:19888
+    # HA support by specifying multiple clusters
+    # e.g.
 
-      # URL of the NodeManager API
-      node_manager_api_url=http://localhost:8042
+    # [[[ha]]]
+      # Enter the host on which you are running the failover JobTracker
+      # jobtracker_host=localhost-ha
 
 
 ###########################################################################

+ 50 - 44
desktop/conf/pseudo-distributed.ini.tmpl

@@ -282,7 +282,7 @@
 
   # Have Hue initiated authn requests be signed and provide a certificate.
   ## authn_requests_signed=false
-  
+
   # Have Hue initiated logout requests be signed and provide a certificate.
   ## logout_requests_signed=false
 
@@ -301,13 +301,14 @@
       # Enter the filesystem uri
       fs_defaultfs=hdfs://localhost:8020
 
-      # Use WebHdfs/HttpFs as the communication mechanism. To fallback to
-      # using the Thrift plugin (used in Hue 1.x), this must be uncommented
-      # and explicitly set to the empty value.
-      ## webhdfs_url=
-
+      # Change this if your HDFS cluster is Kerberos-secured
       ## security_enabled=false
 
+      # Use WebHdfs/HttpFs as the communication mechanism.
+      # This should be the web service root URL, such as
+      # http://namenode:50070/webhdfs/v1
+      ## webhdfs_url=
+
       # Settings about this HDFS cluster. If you install HDFS in a
       # different location, you need to set the following.
 
@@ -320,28 +321,28 @@
       # Defaults to $HADOOP_CONF_DIR or /etc/hadoop/conf
       ## hadoop_conf_dir=/etc/hadoop/conf
 
-  # Configuration for MapReduce 0.20 JobTracker (MR1)
+  # Configuration for YARN (MR2)
   # ------------------------------------------------------------------------
-  [[mapred_clusters]]
+  [[yarn_clusters]]
 
     [[[default]]]
-      # Enter the host on which you are running the Hadoop JobTracker
-      jobtracker_host=localhost
-      # The port where the JobTracker IPC listens on
-      jobtracker_port=8021
-      # Thrift plug-in port for the JobTracker
-      ## thrift_port=9290
+      # Enter the host on which you are running the ResourceManager
+      ## resourcemanager_host=localhost
+
+      # The port where the ResourceManager IPC listens on
+      ## resourcemanager_port=8032
+
       # Whether to submit jobs to this cluster
-      ## submit_to=True
+      submit_to=True
 
-      # Change this if your MapReduce cluster is Kerberos-secured
+      # Change this if your YARN cluster is Kerberos-secured
       ## security_enabled=false
 
-      # Settings about this MR1 cluster. If you install MR1 in a
+      # Settings about this MR2 cluster. If you install MR2 in a
       # different location, you need to set the following.
 
-      # Defaults to $HADOOP_MR1_HOME or /usr/lib/hadoop-0.20-mapreduce
-      ## hadoop_mapred_home=/usr/lib/hadoop-0.20-mapreduce
+      # Defaults to $HADOOP_MR2_HOME or /usr/lib/hadoop-mapreduce
+      ## hadoop_mapred_home=/usr/lib/hadoop-mapreduce
 
       # Defaults to $HADOOP_BIN or /usr/bin/hadoop
       ## hadoop_bin=/usr/bin/hadoop
@@ -349,33 +350,43 @@
       # Defaults to $HADOOP_CONF_DIR or /etc/hadoop/conf
       ## hadoop_conf_dir=/etc/hadoop/conf
 
-    # HA support by specifying multiple clusters
-    # e.g.
+      # URL of the ResourceManager API
+      ## resourcemanager_api_url=http://localhost:8088
 
-    # [[[ha]]]
-      # Enter the host on which you are running the failover JobTracker
-      # jobtracker_host=localhost-ha
+      # URL of the ProxyServer API
+      ## proxy_api_url=http://localhost:8088
 
+      # URL of the HistoryServer API
+      # history_server_api_url=http://localhost:19888
 
-  # Configuration for Yarn
+      # URL of the NodeManager API
+      # node_manager_api_url=http://localhost:8042
+
+  # Configuration for MapReduce (MR1)
   # ------------------------------------------------------------------------
-  [[yarn_clusters]]
+  [[mapred_clusters]]
 
     [[[default]]]
-      # Enter the host on which you are running the ResourceManager
-      resourcemanager_host=localhost
-      # The port where the ResourceManager IPC listens on
-      resourcemanager_port=8032
+      # Enter the host on which you are running the Hadoop JobTracker
+      ## jobtracker_host=localhost
+
+      # The port where the JobTracker IPC listens on
+      ## jobtracker_port=8021
+
+      # Thrift plug-in port for the JobTracker
+      ## thrift_port=9290
+
       # Whether to submit jobs to this cluster
-      ## submit_to=False
+      submit_to=False
 
+      # Change this if your MapReduce cluster is Kerberos-secured
       ## security_enabled=false
 
-      # Settings about this MR2 cluster. If you install MR2 in a
+      # Settings about this MR1 cluster. If you install MR1 in a
       # different location, you need to set the following.
 
-      # Defaults to $HADOOP_MR2_HOME or /usr/lib/hadoop-mapreduce
-      ## hadoop_mapred_home=/usr/lib/hadoop-mapreduce
+      # Defaults to $HADOOP_MR1_HOME or /usr/lib/hadoop-0.20-mapreduce
+      ## hadoop_mapred_home=/usr/lib/hadoop-0.20-mapreduce
 
       # Defaults to $HADOOP_BIN or /usr/bin/hadoop
       ## hadoop_bin=/usr/bin/hadoop
@@ -383,17 +394,12 @@
       # Defaults to $HADOOP_CONF_DIR or /etc/hadoop/conf
       ## hadoop_conf_dir=/etc/hadoop/conf
 
-      # URL of the ResourceManager API
-      ## resourcemanager_api_url=http://localhost:8088
-
-      # URL of the ProxyServer API
-      ## proxy_api_url=http://localhost:8088
-
-      # URL of the HistoryServer API
-      history_server_api_url=http://localhost:19888
+    # HA support by specifying multiple clusters
+    # e.g.
 
-      # URL of the NodeManager API
-      node_manager_api_url=http://localhost:8042
+    # [[[ha]]]
+      # Enter the host on which you are running the failover JobTracker
+      # jobtracker_host=localhost-ha
 
 
 ###########################################################################

+ 2 - 2
desktop/libs/hadoop/src/hadoop/conf.py

@@ -141,7 +141,7 @@ MR_CLUSTERS = UnspecifiedConfigSection(
       SECURITY_ENABLED=Config("security_enabled", help="Is running with Kerberos authentication",
                               default=False, type=coerce_bool),
       SUBMIT_TO=Config('submit_to', help="Whether Hue should use this cluster to run jobs",
-                       default=True, type=coerce_bool),
+                       default=True, type=coerce_bool), # Backward compatibility
 
       HADOOP_MAPRED_HOME = Config(
         key="hadoop_mapred_home",
@@ -185,7 +185,7 @@ YARN_CLUSTERS = UnspecifiedConfigSection(
       SECURITY_ENABLED=Config("security_enabled", help="Is running with Kerberos authentication",
                               default=False, type=coerce_bool),
       SUBMIT_TO=Config('submit_to', help="Whether Hue should use this cluster to run jobs",
-                       default=False, type=coerce_bool),
+                       default=False, type=coerce_bool), # Backward compatibility
 
       HADOOP_MAPRED_HOME = Config(
         key="hadoop_mapred_home",