pseudo-distributed.ini.tmpl 18 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548
  1. #####################################
  2. # DEVELOPMENT EDITION
  3. #####################################
  4. # Hue configuration file
  5. # ===================================
  6. #
  7. # For complete documentation about the contents of this file, run
  8. # $ <hue_root>/build/env/bin/hue config_help
  9. #
  10. # All .ini files under the current directory are treated equally. Their
  11. # contents are merged to form the Hue configuration, which can
  12. # can be viewed on the Hue at
  13. # http://<hue_host>:<port>/dump_config
  14. ###########################################################################
  15. # General configuration for core Desktop features (authentication, etc)
  16. ###########################################################################
  17. [desktop]
  18. send_dbug_messages=1
  19. # To show database transactions, set database_logging to 1
  20. database_logging=0
  21. # Set this to a random string, the longer the better.
  22. # This is used for secure hashing in the session store.
  23. secret_key=
  24. # Webserver listens on this address and port
  25. http_host=0.0.0.0
  26. http_port=8000
  27. # Time zone name
  28. time_zone=America/Los_Angeles
  29. # Turn off debug
  30. django_debug_mode=1
  31. # Turn off backtrace for server error
  32. http_500_debug_mode=1
  33. # Server email for internal error messages
  34. ## django_server_email='hue@localhost.localdomain'
  35. # Email backend
  36. ## django_email_backend=django.core.mail.backends.smtp.EmailBackend
  37. # Set to true to use CherryPy as the webserver, set to false
  38. # to use Spawning as the webserver. Defaults to Spawning if
  39. # key is not specified.
  40. ## use_cherrypy_server = false
  41. # Webserver runs as this user
  42. ## server_user=hue
  43. ## server_group=hue
  44. # If set to false, runcpserver will not actually start the web server.
  45. # Used if Apache is being used as a WSGI container.
  46. ## enable_server=yes
  47. # Number of threads used by the CherryPy web server
  48. ## cherrypy_server_threads=10
  49. # Filename of SSL Certificate
  50. ## ssl_certificate=
  51. # Filename of SSL RSA Private Key
  52. ## ssl_private_key=
  53. # Default encoding for site data
  54. ## default_site_encoding=utf-8
  55. # Administrators
  56. # ----------------
  57. [[django_admins]]
  58. ## [[[admin1]]]
  59. ## name=john
  60. ## email=john@doe.com
  61. # UI customizations
  62. # -------------------
  63. [[custom]]
  64. # Top banner HTML code
  65. ## banner_top_html=
  66. # Configuration options for user authentication into the web application
  67. # ------------------------------------------------------------------------
  68. [[auth]]
  69. # Authentication backend. Common settings are:
  70. # - django.contrib.auth.backends.ModelBackend (entirely Django backend)
  71. # - desktop.auth.backend.AllowAllBackend (allows everyone)
  72. # - desktop.auth.backend.AllowFirstUserDjangoBackend
  73. # (Default. Relies on Django and user manager, after the first login)
  74. # - desktop.auth.backend.LdapBackend
  75. # - desktop.auth.backend.PamBackend
  76. # - desktop.auth.backend.SpnegoDjangoBackend
  77. # - desktop.auth.backend.RemoteUserDjangoBackend
  78. # - desktop.auth.backend.OAuthBackend
  79. ## backend=desktop.auth.backend.AllowFirstUserDjangoBackend
  80. ## pam_service=login
  81. # When using the desktop.auth.backend.RemoteUserDjangoBackend, this sets
  82. # the normalized name of the header that contains the remote user.
  83. # The HTTP header in the request is converted to a key by converting
  84. # all characters to uppercase, replacing any hyphens with underscores
  85. # and adding an HTTP_ prefix to the name. So, for example, if the header
  86. # is called Remote-User that would be configured as HTTP_REMOTE_USER
  87. #
  88. # Defaults to HTTP_REMOTE_USER
  89. ## remote_user_header=HTTP_REMOTE_USER
  90. # Configuration options for connecting to LDAP and Active Directory
  91. # -------------------------------------------------------------------
  92. [[ldap]]
  93. # The search base for finding users and groups
  94. ## base_dn="DC=mycompany,DC=com"
  95. # The NT domain to connect to (only for use with Active Directory)
  96. ## nt_domain=mycompany.com
  97. # URL of the LDAP server
  98. ## ldap_url=ldap://auth.mycompany.com
  99. # A PEM-format file containing certificates for the CA's that
  100. # Hue will trust for authentication over TLS.
  101. # The certificate for the CA that signed the
  102. # LDAP server certificate must be included among these certificates.
  103. # See more here http://www.openldap.org/doc/admin24/tls.html.
  104. ## ldap_cert=
  105. ## use_start_tls=true
  106. # Distinguished name of the user to bind as -- not necessary if the LDAP server
  107. # supports anonymous searches
  108. ## bind_dn="CN=ServiceAccount,DC=mycompany,DC=com"
  109. # Password of the bind user -- not necessary if the LDAP server supports
  110. # anonymous searches
  111. ## bind_password=
  112. # Pattern for searching for usernames -- Use <username> for the parameter
  113. # For use when using LdapBackend for Hue authentication
  114. ## ldap_username_pattern="uid=<username>,ou=People,dc=mycompany,dc=com"
  115. # Create users in Hue when they try to login with their LDAP credentials
  116. # For use when using LdapBackend for Hue authentication
  117. ## create_users_on_login = true
  118. [[[users]]]
  119. # Base filter for searching for users
  120. ## user_filter="objectclass=*"
  121. # The username attribute in the LDAP schema
  122. ## user_name_attr=sAMAccountName
  123. [[[groups]]]
  124. # Base filter for searching for groups
  125. ## group_filter="objectclass=*"
  126. # The username attribute in the LDAP schema
  127. ## group_name_attr=cn
  128. # Configuration options for specifying the Desktop Database. For more info,
  129. # see http://docs.djangoproject.com/en/1.1/ref/settings/#database-engine
  130. # ------------------------------------------------------------------------
  131. [[database]]
  132. # Database engine is typically one of:
  133. # postgresql_psycopg2, mysql, or sqlite3
  134. #
  135. # Note that for sqlite3, 'name', below is a filename;
  136. # for other backends, it is the database name.
  137. ## engine=sqlite3
  138. ## host=
  139. ## port=
  140. ## user=
  141. ## password=
  142. ## name=
  143. # Configuration options for connecting to an external SMTP server
  144. # ------------------------------------------------------------------------
  145. [[smtp]]
  146. # The SMTP server information for email notification delivery
  147. host=localhost
  148. port=25
  149. user=
  150. password=
  151. # Whether to use a TLS (secure) connection when talking to the SMTP server
  152. tls=no
  153. # Default email address to use for various automated notification from Hue
  154. ## default_from_email=hue@localhost
  155. # Configuration options for Kerberos integration for secured Hadoop clusters
  156. # ------------------------------------------------------------------------
  157. [[kerberos]]
  158. # Path to Hue's Kerberos keytab file
  159. ## hue_keytab=
  160. # Kerberos principal name for Hue
  161. ## hue_principal=hue/hostname.foo.com
  162. # Path to kinit
  163. ## kinit_path=/path/to/kinit
  164. # Configuration options for using OAuthBackend login
  165. # ------------------------------------------------------------------------
  166. [[oauth]]
  167. # The Consumer key of the application
  168. ## consumer_key=XXXXXXXXXXXXXXXXXXXXX
  169. # The Consumer secret of the application
  170. ## consumer_secret=XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX
  171. # The Request token URL
  172. ## request_token_url=https://api.twitter.com/oauth/request_token
  173. # The Access token URL
  174. ## access_token_url=https://api.twitter.com/oauth/access_token
  175. # The Authorize URL
  176. ## authenticate_url=https://api.twitter.com/oauth/authorize
  177. ###########################################################################
  178. # Settings to configure your Hadoop cluster.
  179. ###########################################################################
  180. [hadoop]
  181. # Configuration for HDFS NameNode
  182. # ------------------------------------------------------------------------
  183. [[hdfs_clusters]]
  184. [[[default]]]
  185. # Enter the filesystem uri
  186. fs_defaultfs=hdfs://localhost:8020
  187. # Use WebHdfs/HttpFs as the communication mechanism. To fallback to
  188. # using the Thrift plugin (used in Hue 1.x), this must be uncommented
  189. # and explicitly set to the empty value.
  190. ## webhdfs_url=
  191. ## security_enabled=false
  192. # Settings about this HDFS cluster. If you install HDFS in a
  193. # different location, you need to set the following.
  194. # Defaults to $HADOOP_HDFS_HOME or /usr/lib/hadoop-hdfs
  195. ## hadoop_hdfs_home=/usr/lib/hadoop-hdfs
  196. # Defaults to $HADOOP_BIN or /usr/bin/hadoop
  197. ## hadoop_bin=/usr/bin/hadoop
  198. # Defaults to $HADOOP_CONF_DIR or /etc/hadoop/conf
  199. ## hadoop_conf_dir=/etc/hadoop/conf
  200. # Configuration for MapReduce JobTracker
  201. # ------------------------------------------------------------------------
  202. [[mapred_clusters]]
  203. [[[default]]]
  204. # Enter the host on which you are running the Hadoop JobTracker
  205. jobtracker_host=localhost
  206. # The port where the JobTracker IPC listens on
  207. jobtracker_port=8021
  208. # Thrift plug-in port for the JobTracker
  209. ## thrift_port=9290
  210. # Whether to submit jobs to this cluster
  211. ## submit_to=False
  212. ## security_enabled=false
  213. # Settings about this MR1 cluster. If you install MR1 in a
  214. # different location, you need to set the following.
  215. # Defaults to $HADOOP_MR1_HOME or /usr/lib/hadoop-0.20-mapreduce
  216. ## hadoop_mapred_home=/usr/lib/hadoop-0.20-mapreduce
  217. # Defaults to $HADOOP_BIN or /usr/bin/hadoop
  218. ## hadoop_bin=/usr/bin/hadoop
  219. # Defaults to $HADOOP_CONF_DIR or /etc/hadoop/conf
  220. ## hadoop_conf_dir=/etc/hadoop/conf
  221. # Configuration for Yarn
  222. # ------------------------------------------------------------------------
  223. [[yarn_clusters]]
  224. [[[default]]]
  225. # Enter the host on which you are running the ResourceManager
  226. resourcemanager_host=localhost
  227. # The port where the ResourceManager IPC listens on
  228. resourcemanager_port=8032
  229. # Whether to submit jobs to this cluster
  230. ## submit_to=False
  231. ## security_enabled=false
  232. # Settings about this MR2 cluster. If you install MR2 in a
  233. # different location, you need to set the following.
  234. # Defaults to $HADOOP_MR2_HOME or /usr/lib/hadoop-mapreduce
  235. ## hadoop_mapred_home=/usr/lib/hadoop-mapreduce
  236. # Defaults to $HADOOP_BIN or /usr/bin/hadoop
  237. ## hadoop_bin=/usr/bin/hadoop
  238. # Defaults to $HADOOP_CONF_DIR or /etc/hadoop/conf
  239. ## hadoop_conf_dir=/etc/hadoop/conf
  240. # URL of the ResourceManager API
  241. ## resourcemanager_api_url=http://localhost:8088
  242. # URL of the ProxyServer API
  243. ## proxy_api_url=http://localhost:8088
  244. # URL of the HistoryServer API
  245. history_server_api_url=http://localhost:19888
  246. # URL of the NodeManager API
  247. node_manager_api_url=http://localhost:8042
  248. ###########################################################################
  249. # Settings to configure liboozie
  250. ###########################################################################
  251. [liboozie]
  252. # The URL where the Oozie service runs on. This is required in order for
  253. # users to submit jobs.
  254. ## oozie_url=http://localhost:11000/oozie
  255. ## security_enabled=false
  256. # Location on HDFS where the workflows/coordinator are deployed when submitted.
  257. ## remote_deployement_dir=/user/hue/oozie/deployments
  258. ###########################################################################
  259. # Settings to configure the Oozie app
  260. ###########################################################################
  261. [oozie]
  262. # Location on local FS where the examples are stored.
  263. ## local_data_dir=..../examples
  264. # Location on local FS where the data for the examples is stored.
  265. ## sample_data_dir=...thirdparty/sample_data
  266. # Location on HDFS where the oozie examples and workflows are stored.
  267. ## remote_data_dir=/user/hue/oozie/workspaces
  268. # Share workflows and coordinators information with all users. If set to false,
  269. # they will be visible only to the owner and administrators.
  270. ## share_jobs=True
  271. # Maximum of Oozie workflows or coodinators to retrieve in one API call.
  272. ## oozie_jobs_count=100
  273. ###########################################################################
  274. # Settings to configure Beeswax
  275. ###########################################################################
  276. [beeswax]
  277. # Host where Beeswax server Thrift daemon is running.
  278. # If Kerberos security is enabled, the fully-qualified domain name (FQDN) is
  279. # required, even if the Thrift daemon is running on the same host as Hue.
  280. ## beeswax_server_host=<FQDN of Beeswax Server>
  281. # Port where Beeswax Thrift server runs on.
  282. ## beeswax_server_port=8002
  283. # Host where internal metastore Thrift daemon is running.
  284. ## beeswax_meta_server_host=localhost
  285. # Configure the port the internal metastore daemon runs on.
  286. # Used only if hive.metastore.local is true.
  287. ## beeswax_meta_server_port=8003
  288. # Hive home directory
  289. ## hive_home_dir=/usr/lib/hive
  290. # Hive configuration directory, where hive-site.xml is located
  291. ## hive_conf_dir=/etc/hive/conf
  292. # Timeout in seconds for thrift calls to beeswax service
  293. ## beeswax_server_conn_timeout=120
  294. # Timeout in seconds for thrift calls to the hive metastore
  295. ## metastore_conn_timeout=10
  296. # Maximum Java heapsize (in megabytes) used by Beeswax Server.
  297. # Note that the setting of HADOOP_HEAPSIZE in $HADOOP_CONF_DIR/hadoop-env.sh
  298. # may override this setting.
  299. ## beeswax_server_heapsize=1000
  300. # Share saved queries with all users. If set to false, saved queries are
  301. # visible only to the owner and administrators.
  302. ## share_saved_queries=true
  303. # The backend to contact for queries/metadata requests
  304. # Choices are 'beeswax' (default), 'hiveserver2'.
  305. ## server_interface=beeswax
  306. # Time in milliseconds for Beeswax to persist queries in its cache.
  307. # 7*24*60*60*1000 = 1 week
  308. ## beeswax_running_query_lifetime=604800000L
  309. ###########################################################################
  310. # Settings to configure Impala
  311. ###########################################################################
  312. [impala]
  313. # Host of the Impala Server
  314. ## server_host=localhost
  315. # Port of the Impala Server when using Beeswax Thrift interface
  316. ## server_port=21000
  317. # Port of the Impala Server when using Hive Server 2 Thrift interface
  318. ## server_port=21050
  319. # Kerberos principal
  320. ## impala_principal=impala/hostname.foo.com
  321. ###########################################################################
  322. # Settings to configure Job Designer
  323. ###########################################################################
  324. [jobsub]
  325. # Location on HDFS where the jobsub examples and templates are stored.
  326. ## remote_data_dir=/user/hue/jobsub
  327. # Location on local FS where examples and template are stored.
  328. ## local_data_dir=..../data
  329. # Location on local FS where sample data is stored
  330. ## sample_data_dir=...thirdparty/sample_data
  331. ###########################################################################
  332. # Settings to configure Job Browser
  333. ###########################################################################
  334. [jobbrowser]
  335. # Share submitted jobs information with all users. If set to false,
  336. # submitted jobs are visible only to the owner and administrators.
  337. ## share_jobs=true
  338. ###########################################################################
  339. # Settings to configure the Shell application
  340. ###########################################################################
  341. [shell]
  342. # The shell_buffer_amount specifies the number of bytes of output per shell
  343. # that the Shell app will keep in memory. If not specified, it defaults to
  344. # 524288 (512 MiB).
  345. ## shell_buffer_amount=100
  346. # If you run Hue against a Hadoop cluster with Kerberos security enabled, the
  347. # Shell app needs to acquire delegation tokens for the subprocesses to work
  348. # correctly. These delegation tokens are stored as temporary files in some
  349. # directory. You can configure this directory here. If not specified, it
  350. # defaults to /tmp/hue_delegation_tokens.
  351. ## shell_delegation_token_dir=/tmp/hue_delegation_tokens
  352. [[ shelltypes ]]
  353. # Define and configure a new shell type "flume"
  354. # ------------------------------------------------------------------------
  355. [[[ flume ]]]
  356. nice_name = "Flume Shell"
  357. command = "/usr/bin/flume shell"
  358. help = "The command-line Flume client interface."
  359. [[[[ environment ]]]]
  360. # You can specify environment variables for the Flume shell
  361. # in this section.
  362. # Define and configure a new shell type "pig"
  363. # ------------------------------------------------------------------------
  364. [[[ pig ]]]
  365. nice_name = "Pig Shell (Grunt)"
  366. command = "/usr/bin/pig -l /dev/null"
  367. help = "The command-line interpreter for Pig"
  368. [[[[ environment ]]]]
  369. # You can specify environment variables for the Pig shell
  370. # in this section. Note that JAVA_HOME must be configured
  371. # for the Pig shell to run.
  372. [[[[[ JAVA_HOME ]]]]]
  373. value = "/usr/lib/jvm/java-6-sun"
  374. # Define and configure a new shell type "hbase"
  375. # ------------------------------------------------------------------------
  376. [[[ hbase ]]]
  377. nice_name = "HBase Shell"
  378. command = "/usr/bin/hbase shell"
  379. help = "The command-line HBase client interface."
  380. [[[[ environment ]]]]
  381. # You can configure environment variables for the HBase shell
  382. # in this section.
  383. # Define and configure a new shell type "Sqoop 2"
  384. # ------------------------------------------------------------------------
  385. [[[ sqoop2 ]]]
  386. nice_name = "Sqoop 2 Shell"
  387. command = "/usr/bin/sqoop2"
  388. help = "The command-line Sqoop 2 client."
  389. [[[[ environment ]]]]
  390. # You can configure environment variables for the Sqoop 2 shell
  391. # in this section.
  392. # Define and configure a new shell type "bash" for testing only
  393. # ------------------------------------------------------------------------
  394. [[[ bash ]]]
  395. nice_name = "Bash (Test only!!!)"
  396. command = "/bin/bash"
  397. help = "A shell that does not depend on Hadoop components"
  398. ###########################################################################
  399. # Settings for the User Admin application
  400. ###########################################################################
  401. [useradmin]
  402. # The name of the default user group that users will be a member of
  403. ## default_user_group=default