hue.ini 20 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613
  1. # Hue configuration file
  2. # ===================================
  3. #
  4. # For complete documentation about the contents of this file, run
  5. # $ <hue_root>/build/env/bin/hue config_help
  6. #
  7. # All .ini files under the current directory are treated equally. Their
  8. # contents are merged to form the Hue configuration, which can
  9. # can be viewed on the Hue at
  10. # http://<hue_host>:<port>/dump_config
  11. ###########################################################################
  12. # General configuration for core Desktop features (authentication, etc)
  13. ###########################################################################
  14. [desktop]
  15. # Set this to a random string, the longer the better.
  16. # This is used for secure hashing in the session store.
  17. secret_key=
  18. # Webserver listens on this address and port
  19. http_host=0.0.0.0
  20. http_port=8888
  21. # Time zone name
  22. time_zone=America/Los_Angeles
  23. # Turn off debug
  24. django_debug_mode=0
  25. # Turn off backtrace for server error
  26. http_500_debug_mode=0
  27. # Server email for internal error messages
  28. ## django_server_email='hue@localhost.localdomain'
  29. # Email backend
  30. ## django_email_backend=django.core.mail.backends.smtp.EmailBackend
  31. # Set to true to use CherryPy as the webserver, set to false
  32. # to use Spawning as the webserver. Defaults to Spawning if
  33. # key is not specified.
  34. ## use_cherrypy_server = false
  35. # Webserver runs as this user
  36. ## server_user=hue
  37. ## server_group=hue
  38. # If set to false, runcpserver will not actually start the web server.
  39. # Used if Apache is being used as a WSGI container.
  40. ## enable_server=yes
  41. # Number of threads used by the CherryPy web server
  42. ## cherrypy_server_threads=10
  43. # Filename of SSL Certificate
  44. ## ssl_certificate=
  45. # Filename of SSL RSA Private Key
  46. ## ssl_private_key=
  47. # Default encoding for site data
  48. ## default_site_encoding=utf-8
  49. # Help improve Hue with anonymous usage analytics.
  50. # Use Google Analytics to see how many times an application or specific section of an application is used, nothing more.
  51. ## collect_usage=true
  52. # Administrators
  53. # ----------------
  54. [[django_admins]]
  55. ## [[[admin1]]]
  56. ## name=john
  57. ## email=john@doe.com
  58. # UI customizations
  59. # -------------------
  60. [[custom]]
  61. # Top banner HTML code
  62. ## banner_top_html=
  63. # Configuration options for user authentication into the web application
  64. # ------------------------------------------------------------------------
  65. [[auth]]
  66. # Authentication backend. Common settings are:
  67. # - django.contrib.auth.backends.ModelBackend (entirely Django backend)
  68. # - desktop.auth.backend.AllowAllBackend (allows everyone)
  69. # - desktop.auth.backend.AllowFirstUserDjangoBackend
  70. # (Default. Relies on Django and user manager, after the first login)
  71. # - desktop.auth.backend.LdapBackend
  72. # - desktop.auth.backend.PamBackend
  73. # - desktop.auth.backend.SpnegoDjangoBackend
  74. # - desktop.auth.backend.RemoteUserDjangoBackend
  75. # - desktop.auth.backend.OAuthBackend
  76. ## backend=desktop.auth.backend.AllowFirstUserDjangoBackend
  77. # Backend to synchronize user-group membership with
  78. ## user_group_membership_synchronization_backend=desktop.auth.backend.LdapSynchronizationBackend
  79. ## pam_service=login
  80. # When using the desktop.auth.backend.RemoteUserDjangoBackend, this sets
  81. # the normalized name of the header that contains the remote user.
  82. # The HTTP header in the request is converted to a key by converting
  83. # all characters to uppercase, replacing any hyphens with underscores
  84. # and adding an HTTP_ prefix to the name. So, for example, if the header
  85. # is called Remote-User that would be configured as HTTP_REMOTE_USER
  86. #
  87. # Defaults to HTTP_REMOTE_USER
  88. ## remote_user_header=HTTP_REMOTE_USER
  89. # Configuration options for connecting to LDAP and Active Directory
  90. # -------------------------------------------------------------------
  91. [[ldap]]
  92. # The search base for finding users and groups
  93. ## base_dn="DC=mycompany,DC=com"
  94. # URL of the LDAP server
  95. ## ldap_url=ldap://auth.mycompany.com
  96. # A PEM-format file containing certificates for the CA's that
  97. # Hue will trust for authentication over TLS.
  98. # The certificate for the CA that signed the
  99. # LDAP server certificate must be included among these certificates.
  100. # See more here http://www.openldap.org/doc/admin24/tls.html.
  101. ## ldap_cert=
  102. ## use_start_tls=true
  103. # Distinguished name of the user to bind as -- not necessary if the LDAP server
  104. # supports anonymous searches
  105. ## bind_dn="CN=ServiceAccount,DC=mycompany,DC=com"
  106. # Password of the bind user -- not necessary if the LDAP server supports
  107. # anonymous searches
  108. ## bind_password=
  109. # Create users in Hue when they try to login with their LDAP credentials
  110. # For use when using LdapBackend for Hue authentication
  111. ## create_users_on_login = true
  112. [[[users]]]
  113. # Base filter for searching for users
  114. ## user_filter="objectclass=*"
  115. # The username attribute in the LDAP schema
  116. ## user_name_attr=sAMAccountName
  117. [[[groups]]]
  118. # Base filter for searching for groups
  119. ## group_filter="objectclass=*"
  120. # The group name attribute in the LDAP schema
  121. ## group_name_attr=cn
  122. # The attribute of the group object which identifies the members of the group
  123. ## group_member_attr=members
  124. # Configuration options for specifying the Desktop Database. For more info,
  125. # see http://docs.djangoproject.com/en/1.1/ref/settings/#database-engine
  126. # ------------------------------------------------------------------------
  127. [[database]]
  128. # Database engine is typically one of:
  129. # postgresql_psycopg2, mysql, or sqlite3
  130. #
  131. # Note that for sqlite3, 'name', below is a filename;
  132. # for other backends, it is the database name.
  133. ## engine=sqlite3
  134. ## host=
  135. ## port=
  136. ## user=
  137. ## password=
  138. ## name=desktop/desktop.db
  139. # Configuration options for connecting to an external SMTP server
  140. # ------------------------------------------------------------------------
  141. [[smtp]]
  142. # The SMTP server information for email notification delivery
  143. host=localhost
  144. port=25
  145. user=
  146. password=
  147. # Whether to use a TLS (secure) connection when talking to the SMTP server
  148. tls=no
  149. # Default email address to use for various automated notification from Hue
  150. ## default_from_email=hue@localhost
  151. # Configuration options for Kerberos integration for secured Hadoop clusters
  152. # ------------------------------------------------------------------------
  153. [[kerberos]]
  154. # Path to Hue's Kerberos keytab file
  155. ## hue_keytab=
  156. # Kerberos principal name for Hue
  157. ## hue_principal=hue/hostname.foo.com
  158. # Path to kinit
  159. ## kinit_path=/path/to/kinit
  160. # Configuration options for using OAuthBackend login
  161. # ------------------------------------------------------------------------
  162. [[oauth]]
  163. # The Consumer key of the application
  164. ## consumer_key=XXXXXXXXXXXXXXXXXXXXX
  165. # The Consumer secret of the application
  166. ## consumer_secret=XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX
  167. # The Request token URL
  168. ## request_token_url=https://api.twitter.com/oauth/request_token
  169. # The Access token URL
  170. ## access_token_url=https://api.twitter.com/oauth/access_token
  171. # The Authorize URL
  172. ## authenticate_url=https://api.twitter.com/oauth/authorize
  173. ###########################################################################
  174. # Settings to configure your Hadoop cluster.
  175. ###########################################################################
  176. [hadoop]
  177. # Configuration for HDFS NameNode
  178. # ------------------------------------------------------------------------
  179. [[hdfs_clusters]]
  180. # HA support by using HttpFs
  181. [[[default]]]
  182. # Enter the filesystem uri
  183. fs_defaultfs=hdfs://localhost:8020
  184. # Change this if your HDFS cluster is Kerberos-secured
  185. ## security_enabled=false
  186. # Use WebHdfs/HttpFs as the communication mechanism.
  187. # This should be the web service root URL, such as
  188. # http://namenode:50070/webhdfs/v1
  189. ## webhdfs_url=
  190. # Settings about this HDFS cluster. If you install HDFS in a
  191. # different location, you need to set the following.
  192. # Defaults to $HADOOP_HDFS_HOME or /usr/lib/hadoop-hdfs
  193. ## hadoop_hdfs_home=/usr/lib/hadoop-hdfs
  194. # Defaults to $HADOOP_BIN or /usr/bin/hadoop
  195. ## hadoop_bin=/usr/bin/hadoop
  196. # Defaults to $HADOOP_CONF_DIR or /etc/hadoop/conf
  197. ## hadoop_conf_dir=/etc/hadoop/conf
  198. # Configuration for MapReduce 0.20 JobTracker (MR1)
  199. # ------------------------------------------------------------------------
  200. [[mapred_clusters]]
  201. [[[default]]]
  202. # Enter the host on which you are running the Hadoop JobTracker
  203. jobtracker_host=localhost
  204. # The port where the JobTracker IPC listens on
  205. jobtracker_port=8021
  206. # Thrift plug-in port for the JobTracker
  207. ## thrift_port=9290
  208. # Whether to submit jobs to this cluster
  209. ## submit_to=True
  210. # Change this if your MapReduce cluster is Kerberos-secured
  211. ## security_enabled=false
  212. # Settings about this MR1 cluster. If you install MR1 in a
  213. # different location, you need to set the following.
  214. # Defaults to $HADOOP_MR1_HOME or /usr/lib/hadoop-0.20-mapreduce
  215. ## hadoop_mapred_home=/usr/lib/hadoop-0.20-mapreduce
  216. # Defaults to $HADOOP_BIN or /usr/bin/hadoop
  217. ## hadoop_bin=/usr/bin/hadoop
  218. # Defaults to $HADOOP_CONF_DIR or /etc/hadoop/conf
  219. ## hadoop_conf_dir=/etc/hadoop/conf
  220. # HA support by specifying multiple clusters
  221. # e.g.
  222. # [[[ha]]]
  223. # Enter the host on which you are running the failover JobTracker
  224. # jobtracker_host=localhost-ha
  225. # Configuration for YARN (MR2)
  226. # ------------------------------------------------------------------------
  227. [[yarn_clusters]]
  228. [[[default]]]
  229. # Enter the host on which you are running the ResourceManager
  230. resourcemanager_host=localhost
  231. # The port where the ResourceManager IPC listens on
  232. resourcemanager_port=8032
  233. # Whether to submit jobs to this cluster
  234. ## submit_to=False
  235. # Change this if your YARN cluster is Kerberos-secured
  236. ## security_enabled=false
  237. # Settings about this MR2 cluster. If you install MR2 in a
  238. # different location, you need to set the following.
  239. # Defaults to $HADOOP_MR2_HOME or /usr/lib/hadoop-mapreduce
  240. ## hadoop_mapred_home=/usr/lib/hadoop-mapreduce
  241. # Defaults to $HADOOP_BIN or /usr/bin/hadoop
  242. ## hadoop_bin=/usr/bin/hadoop
  243. # Defaults to $HADOOP_CONF_DIR or /etc/hadoop/conf
  244. ## hadoop_conf_dir=/etc/hadoop/conf
  245. # URL of the ResourceManager API
  246. ## resourcemanager_api_url=http://localhost:8088
  247. # URL of the ProxyServer API
  248. ## proxy_api_url=http://localhost:8088
  249. # URL of the HistoryServer API
  250. history_server_api_url=http://localhost:19888
  251. # URL of the NodeManager API
  252. node_manager_api_url=http://localhost:8042
  253. ###########################################################################
  254. # Settings to configure liboozie
  255. ###########################################################################
  256. [liboozie]
  257. # The URL where the Oozie service runs on. This is required in order for
  258. # users to submit jobs.
  259. ## oozie_url=http://localhost:11000/oozie
  260. # Requires FQDN in oozie_url if enabled
  261. ## security_enabled=false
  262. # Location on HDFS where the workflows/coordinator are deployed when submitted.
  263. ## remote_deployement_dir=/user/hue/oozie/deployments
  264. ###########################################################################
  265. # Settings to configure the Oozie app
  266. ###########################################################################
  267. [oozie]
  268. # Location on local FS where the examples are stored.
  269. ## local_data_dir=..../examples
  270. # Location on local FS where the data for the examples is stored.
  271. ## sample_data_dir=...thirdparty/sample_data
  272. # Location on HDFS where the oozie examples and workflows are stored.
  273. ## remote_data_dir=/user/hue/oozie/workspaces
  274. # Share workflows and coordinators information with all users. If set to false,
  275. # they will be visible only to the owner and administrators.
  276. ## share_jobs=True
  277. # Maximum of Oozie workflows or coodinators to retrieve in one API call.
  278. ## oozie_jobs_count=100
  279. ###########################################################################
  280. # Settings to configure Beeswax
  281. ###########################################################################
  282. [beeswax]
  283. # Host where Beeswax server Thrift daemon is running.
  284. # If Kerberos security is enabled, the fully-qualified domain name (FQDN) is
  285. # required, even if the Thrift daemon is running on the same host as Hue.
  286. ## beeswax_server_host=<FQDN of Beeswax Server>
  287. # The type of Thrift interface used for contacting the backend for sending
  288. # queries/metadata requests.
  289. # Choices are 'beeswax' (default), 'hiveserver2'.
  290. ## server_interface=beeswax
  291. # Port where Beeswax Thrift server runs on.
  292. # Use 10000 when using the HiveServer2 interface.
  293. ## beeswax_server_port=8002
  294. # Host where internal metastore Thrift daemon is running.
  295. ## beeswax_meta_server_host=localhost
  296. # Configure the port the internal metastore daemon runs on.
  297. # Used only if hive.metastore.local is true.
  298. ## beeswax_meta_server_port=8003
  299. # Hive home directory
  300. ## hive_home_dir=/usr/lib/hive
  301. # Hive configuration directory, where hive-site.xml is located
  302. ## hive_conf_dir=/etc/hive/conf
  303. # Timeout in seconds for thrift calls to beeswax service
  304. ## beeswax_server_conn_timeout=120
  305. # Timeout in seconds for thrift calls to the hive metastore
  306. ## metastore_conn_timeout=10
  307. # Maximum Java heapsize (in megabytes) used by Beeswax Server.
  308. # Note that the setting of HADOOP_HEAPSIZE in $HADOOP_CONF_DIR/hadoop-env.sh
  309. # may override this setting.
  310. ## beeswax_server_heapsize=1000
  311. # Share saved queries with all users. If set to false, saved queries are
  312. # visible only to the owner and administrators.
  313. ## share_saved_queries=true
  314. # Time in milliseconds for Beeswax to persist queries in its cache.
  315. # 7*24*60*60*1000 = 1 week
  316. ## beeswax_running_query_lifetime=604800000L
  317. # Set a LIMIT clause when browsing a partitioned table.
  318. # A positive value will be set as the LIMIT. If 0 or negative, do not set any limit.
  319. ## browse_partitioned_table_limit=250
  320. ###########################################################################
  321. # Settings to configure Pig
  322. ###########################################################################
  323. [pig]
  324. # Location of piggybank.jar on local filesystem.
  325. ## local_sample_dir=/usr/share/hue/apps/pig/examples
  326. # Location piggybank.jar will be copied to in HDFS.
  327. ## remote_data_dir=/user/hue/pig/examples
  328. ###########################################################################
  329. # Settings to configure Sqoop
  330. ###########################################################################
  331. [sqoop]
  332. # Sqoop server URL
  333. ## server_url=http://localhost:12000/sqoop
  334. ###########################################################################
  335. # Settings to configure Proxy
  336. ###########################################################################
  337. [proxy]
  338. # Comma-separated list of regular expressions,
  339. # which match 'host:port' of requested proxy target.
  340. ## whitelist=(localhost|127\.0\.0\.1):(50030|50070|50060|50075)
  341. # Comma-separated list of regular expressions,
  342. # which match any prefix of 'host:port/path' of requested proxy target.
  343. # This does not support matching GET parameters.
  344. ## blacklist=()
  345. ###########################################################################
  346. # Settings to configure Impala
  347. ###########################################################################
  348. [impala]
  349. # Host of the Impala Server (one of the Impalad)
  350. ## server_host=localhost
  351. # The backend to contact for queries/metadata requests.
  352. # Choices are 'beeswax' or 'hiveserver2' (default).
  353. # 'hiveserver2' supports log, progress information, query cancelation
  354. # 'beeswax' requires Beeswax to run for proxying the metadata requests
  355. ## server_interface=hiveserver2
  356. # Port of the Impala Server
  357. # Default is 21050 as HiveServer2 Thrift interface is the default.
  358. # Use 21000 when using Beeswax Thrift interface.
  359. ## server_port=21050
  360. # Kerberos principal
  361. ## impala_principal=impala/hostname.foo.com
  362. ###########################################################################
  363. # Settings to configure Hbase
  364. ###########################################################################
  365. [hbase]
  366. # Comma-separated list of HBase Thrift servers for
  367. # clusters in the format of '(name|host:port)'.
  368. ## hbase_clusters=(Cluster|localhost:9090)
  369. # Hard limit of rows or columns per row fetched before truncating.
  370. ## truncate_limit = 500
  371. ###########################################################################
  372. # Settings to configure Solr Search
  373. ###########################################################################
  374. [search]
  375. # URL of the Solr Server
  376. ## solr_url=http://localhost:8983/solr/
  377. # Requires FQDN in solr_url if enabled
  378. ## security_enabled=false
  379. ## Query sent when no term is entered
  380. ## empty_query=*:*
  381. ###########################################################################
  382. # Settings to configure Job Designer
  383. ###########################################################################
  384. [jobsub]
  385. # Location on HDFS where the jobsub examples and templates are stored.
  386. ## remote_data_dir=/user/hue/jobsub
  387. # Location on local FS where examples and template are stored.
  388. ## local_data_dir=..../data
  389. # Location on local FS where sample data is stored
  390. ## sample_data_dir=...thirdparty/sample_data
  391. ###########################################################################
  392. # Settings to configure Job Browser.
  393. ###########################################################################
  394. [jobbrowser]
  395. # Share submitted jobs information with all users. If set to false,
  396. # submitted jobs are visible only to the owner and administrators.
  397. ## share_jobs=true
  398. ###########################################################################
  399. # Settings to configure the Shell application
  400. ###########################################################################
  401. [shell]
  402. # The shell_buffer_amount specifies the number of bytes of output per shell
  403. # that the Shell app will keep in memory. If not specified, it defaults to
  404. # 524288 (512 MiB).
  405. ## shell_buffer_amount=100
  406. # If you run Hue against a Hadoop cluster with Kerberos security enabled, the
  407. # Shell app needs to acquire delegation tokens for the subprocesses to work
  408. # correctly. These delegation tokens are stored as temporary files in some
  409. # directory. You can configure this directory here. If not specified, it
  410. # defaults to /tmp/hue_delegation_tokens.
  411. ## shell_delegation_token_dir=/tmp/hue_delegation_tokens
  412. [[ shelltypes ]]
  413. # Define and configure a new shell type "pig"
  414. # ------------------------------------------------------------------------
  415. [[[ pig ]]]
  416. nice_name = "Pig Shell (Grunt)"
  417. command = "/usr/bin/pig -l /dev/null"
  418. help = "The command-line interpreter for Pig"
  419. [[[[ environment ]]]]
  420. # You can specify environment variables for the Pig shell
  421. # in this section. Note that JAVA_HOME must be configured
  422. # for the Pig shell to run.
  423. [[[[[ JAVA_HOME ]]]]]
  424. value = "/usr/lib/jvm/java-6-sun"
  425. # Define and configure a new shell type "Sqoop 2"
  426. # ------------------------------------------------------------------------
  427. [[[ sqoop2 ]]]
  428. nice_name = "Sqoop 2 Shell"
  429. command = "/usr/bin/sqoop2"
  430. help = "The command-line Sqoop 2 client."
  431. [[[[ environment ]]]]
  432. # You can configure environment variables for the Sqoop 2 shell
  433. # in this section.
  434. # Define and configure a new shell type "hbase"
  435. # ------------------------------------------------------------------------
  436. [[[ hbase ]]]
  437. nice_name = "HBase Shell"
  438. command = "/usr/bin/hbase shell"
  439. help = "The command-line HBase client interface."
  440. [[[[ environment ]]]]
  441. # You can configure environment variables for the HBase shell
  442. # in this section.
  443. ###########################################################################
  444. # Settings for the User Admin application
  445. ###########################################################################
  446. [useradmin]
  447. # The name of the default user group that users will be a member of
  448. ## default_user_group=default