hue.ini 15 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467
  1. # Hue configuration file
  2. # ===================================
  3. #
  4. # For complete documentation about the contents of this file, run
  5. # $ <hue_root>/build/env/bin/hue config_help
  6. #
  7. # All .ini files under the current directory are treated equally. Their
  8. # contents are merged to form the Hue configuration, which can
  9. # can be viewed on the Hue at
  10. # http://<hue_host>:<port>/dump_config
  11. ###########################################################################
  12. # General configuration for core Desktop features (authentication, etc)
  13. ###########################################################################
  14. [desktop]
  15. # Set this to a random string, the longer the better.
  16. # This is used for secure hashing in the session store.
  17. secret_key=
  18. # Webserver listens on this address and port
  19. http_host=0.0.0.0
  20. http_port=8888
  21. # Time zone name
  22. time_zone=America/Los_Angeles
  23. # Turn off debug
  24. django_debug_mode=0
  25. # Turn off backtrace for server error
  26. http_500_debug_mode=0
  27. # Server email for internal error messages
  28. ## django_server_email='hue@localhost.localdomain'
  29. # Email backend
  30. ## django_email_backend=django.core.mail.backends.smtp.EmailBackend
  31. # Set to true to use CherryPy as the webserver, set to false
  32. # to use Spawning as the webserver. Defaults to Spawning if
  33. # key is not specified.
  34. ## use_cherrypy_server = false
  35. # Webserver runs as this user
  36. ## server_user=hue
  37. ## server_group=hue
  38. # If set to false, runcpserver will not actually start the web server.
  39. # Used if Apache is being used as a WSGI container.
  40. ## enable_server=yes
  41. # Number of threads used by the CherryPy web server
  42. ## cherrypy_server_threads=10
  43. # Filename of SSL Certificate
  44. ## ssl_certificate=
  45. # Filename of SSL RSA Private Key
  46. ## ssl_private_key=
  47. # Default encoding for site data
  48. ## default_site_encoding=utf-8
  49. # Administrators
  50. # ----------------
  51. [[django_admins]]
  52. ## [[[admin1]]]
  53. ## name=john
  54. ## email=john@doe.com
  55. # UI customizations
  56. # -------------------
  57. [[custom]]
  58. # Top banner HTML code
  59. ## banner_top_html=
  60. # Configuration options for user authentication into the web application
  61. # ------------------------------------------------------------------------
  62. [[auth]]
  63. # Authentication backend. Common settings are:
  64. # - django.contrib.auth.backends.ModelBackend (entirely Django backend)
  65. # - desktop.auth.backend.AllowAllBackend (allows everyone)
  66. # - desktop.auth.backend.AllowFirstUserDjangoBackend
  67. # (Default. Relies on Django and user manager, after the first login)
  68. # - desktop.auth.backend.LdapBackend
  69. # - desktop.auth.backend.PamBackend
  70. ## backend=desktop.auth.backend.AllowFirstUserDjangoBackend
  71. ## pam_service=login
  72. # Configuration options for connecting to LDAP and Active Directory
  73. # -------------------------------------------------------------------
  74. [[ldap]]
  75. # The search base for finding users and groups
  76. ## base_dn="DC=mycompany,DC=com"
  77. # The NT domain to connect to (only for use with Active Directory)
  78. ## nt_domain=mycompany.com
  79. # URL of the LDAP server
  80. ## ldap_url=ldap://auth.mycompany.com
  81. # Path to certificate for authentication over TLS
  82. ## ldap_cert=
  83. # Distinguished name of the user to bind as -- not necessary if the LDAP server
  84. # supports anonymous searches
  85. ## bind_dn="CN=ServiceAccount,DC=mycompany,DC=com"
  86. # Password of the bind user -- not necessary if the LDAP server supports
  87. # anonymous searches
  88. ## bind_password=
  89. # Pattern for searching for usernames -- Use <username> for the parameter
  90. # For use when using LdapBackend for Hue authentication
  91. ## ldap_username_pattern="uid=<username>,ou=People,dc=mycompany,dc=com"
  92. [[[users]]]
  93. # Base filter for searching for users
  94. ## user_filter="objectclass=*"
  95. # The username attribute in the LDAP schema
  96. ## user_name_attr=sAMAccountName
  97. [[[groups]]]
  98. # Base filter for searching for groups
  99. ## group_filter="objectclass=*"
  100. # The group name attribute in the LDAP schema
  101. ## group_name_attr=cn
  102. # The attribute of the group object which identifies the members of the group
  103. ## group_member_attr=members
  104. # Configuration options for specifying the Desktop Database. For more info,
  105. # see http://docs.djangoproject.com/en/1.1/ref/settings/#database-engine
  106. # ------------------------------------------------------------------------
  107. [[database]]
  108. # Database engine is typically one of:
  109. # postgresql, mysql, sqlite3, or oracle
  110. #
  111. # Note that for sqlite3, 'name', below is a filename;
  112. # for other backends, it is the database name.
  113. ## engine=sqlite3
  114. ## host=
  115. ## port=
  116. ## user=
  117. ## password=
  118. ## name=
  119. # Configuration options for connecting to an external SMTP server
  120. # ------------------------------------------------------------------------
  121. [[smtp]]
  122. # The SMTP server information for email notification delivery
  123. host=localhost
  124. port=25
  125. user=
  126. password=
  127. # Whether to use a TLS (secure) connection when talking to the SMTP server
  128. tls=no
  129. # Default email address to use for various automated notification from Hue
  130. ## default_from_email=hue@localhost
  131. # Configuration options for Kerberos integration for secured Hadoop clusters
  132. # ------------------------------------------------------------------------
  133. [[kerberos]]
  134. # Path to Hue's Kerberos keytab file
  135. ## hue_keytab=
  136. # Kerberos principal name for Hue
  137. ## hue_principal=hue/hostname.foo.com
  138. # Path to kinit
  139. ## kinit_path=/path/to/kinit
  140. ###########################################################################
  141. # Settings to configure your Hadoop cluster.
  142. ###########################################################################
  143. [hadoop]
  144. # Configuration for HDFS NameNode
  145. # ------------------------------------------------------------------------
  146. [[hdfs_clusters]]
  147. [[[default]]]
  148. # Enter the filesystem uri
  149. fs_defaultfs=hdfs://localhost:8020
  150. # Change this if your HDFS cluster is Kerberos-secured
  151. ## security_enabled=false
  152. # Use WebHdfs/HttpFs as the communication mechanism.
  153. # This should be the web service root URL, such as
  154. # http://namenode:50070/webhdfs/v1
  155. ## webhdfs_url=
  156. # Settings about this HDFS cluster. If you install HDFS in a
  157. # different location, you need to set the following.
  158. # Defaults to $HADOOP_HDFS_HOME or /usr/lib/hadoop-hdfs
  159. ## hadoop_hdfs_home=/usr/lib/hadoop-hdfs
  160. # Defaults to $HADOOP_BIN or /usr/bin/hadoop
  161. ## hadoop_bin=/usr/bin/hadoop
  162. # Defaults to $HADOOP_CONF_DIR or /etc/hadoop/conf
  163. ## hadoop_conf_dir=/etc/hadoop/conf
  164. # Configuration for MapReduce 0.20 JobTracker (MR1)
  165. # ------------------------------------------------------------------------
  166. [[mapred_clusters]]
  167. [[[default]]]
  168. # Enter the host on which you are running the Hadoop JobTracker
  169. jobtracker_host=localhost
  170. # The port where the JobTracker IPC listens on
  171. jobtracker_port=8021
  172. # Thrift plug-in port for the JobTracker
  173. ## thrift_port=9290
  174. # Whether to submit jobs to this cluster
  175. ## submit_to=True
  176. # Change this if your MapReduce cluster is Kerberos-secured
  177. ## security_enabled=false
  178. # Settings about this MR1 cluster. If you install MR1 in a
  179. # different location, you need to set the following.
  180. # Defaults to $HADOOP_MR1_HOME or /usr/lib/hadoop-0.20-mapreduce
  181. ## hadoop_mapred_home=/usr/lib/hadoop-0.20-mapreduce
  182. # Defaults to $HADOOP_BIN or /usr/bin/hadoop
  183. ## hadoop_bin=/usr/bin/hadoop
  184. # Defaults to $HADOOP_CONF_DIR or /etc/hadoop/conf
  185. ## hadoop_conf_dir=/etc/hadoop/conf
  186. # Configuration for YARN (MR2)
  187. # ------------------------------------------------------------------------
  188. [[yarn_clusters]]
  189. [[[default]]]
  190. # Enter the host on which you are running the ResourceManager
  191. resourcemanager_host=localhost
  192. # The port where the ResourceManager IPC listens on
  193. resourcemanager_port=8032
  194. # Whether to submit jobs to this cluster
  195. ## submit_to=False
  196. # Change this if your YARN cluster is Kerberos-secured
  197. ## security_enabled=false
  198. # Settings about this MR2 cluster. If you install MR2 in a
  199. # different location, you need to set the following.
  200. # Defaults to $HADOOP_MR2_HOME or /usr/lib/hadoop-mapreduce
  201. ## hadoop_mapred_home=/usr/lib/hadoop-mapreduce
  202. # Defaults to $HADOOP_BIN or /usr/bin/hadoop
  203. ## hadoop_bin=/usr/bin/hadoop
  204. # Defaults to $HADOOP_CONF_DIR or /etc/hadoop/conf
  205. ## hadoop_conf_dir=/etc/hadoop/conf
  206. ###########################################################################
  207. # Settings to configure liboozie
  208. ###########################################################################
  209. [liboozie]
  210. # The URL where the Oozie service runs on. This is required in order for
  211. # users to submit jobs.
  212. ## oozie_url=http://localhost:11000/oozie
  213. ## security_enabled=false
  214. # Location on HDFS where the workflows/coordinator are deployed when submitted.
  215. ## remote_deployement_dir=/user/hue/oozie/deployments
  216. ###########################################################################
  217. # Settings to configure the Oozie app
  218. ###########################################################################
  219. [oozie]
  220. # Location on local FS where the examples are stored.
  221. ## local_data_dir=..../examples
  222. # Location on local FS where the data for the examples is stored.
  223. ## sample_data_dir=...thirdparty/sample_data
  224. # Location on HDFS where the oozie examples and workflows are stored.
  225. ## remote_data_dir=/user/hue/oozie/workspaces
  226. # Share workflows and coordinators information with all users. If set to false,
  227. # they will be visible only to the owner and administrators.
  228. ## share_jobs=True
  229. # Maximum of Oozie workflows or coodinators to retrieve in one API call.
  230. ## oozie_jobs_count=100
  231. ###########################################################################
  232. # Settings to configure Beeswax
  233. ###########################################################################
  234. [beeswax]
  235. # Deprecated! Will be removed in Hue 3
  236. # Multiple sections are now available in query_servers
  237. # Host where Beeswax internal metastore Thrift daemon is running
  238. ## beeswax_meta_server_host=localhost
  239. # Deprecated! Will be removed in Hue 3
  240. # Multiple sections are now available in query_servers
  241. # Configure the port the internal metastore daemon runs on. Used only if
  242. # hive.metastore.local is true.
  243. ## beeswax_meta_server_port=8003
  244. # Host where Beeswax internal metastore Thrift daemon is running
  245. ## beeswax_meta_server_host=localhost
  246. # Configure the port the internal metastore daemon runs on. Used only if
  247. # hive.metastore.local is true.
  248. ## beeswax_meta_server_port=8003
  249. # Hive home directory
  250. ## hive_home_dir=/usr/lib/hive
  251. # Hive configuration directory, where hive-site.xml is located
  252. ## hive_conf_dir=/etc/hive/conf
  253. # Timeout in seconds for thrift calls to beeswax service
  254. ## beeswax_server_conn_timeout=120
  255. # Timeout in seconds for thrift calls to the hive metastore
  256. ## metastore_conn_timeout=10
  257. # Maximum Java heapsize (in megabytes) used by Beeswax Server.
  258. # Note that the setting of HADOOP_HEAPSIZE in $HADOOP_CONF_DIR/hadoop-env.sh
  259. # may override this setting.
  260. ## beeswax_server_heapsize=1000
  261. # Share saved queries with all users. If set to false, saved queries are
  262. # visible only to the owner and administrators.
  263. ## share_saved_queries=true
  264. # The backend to contact for queries/metadata requests.
  265. # Choices are 'beeswax' (default), 'hiveserver2'.
  266. ## server_interface=beeswax
  267. # One entry for each Query Server that can execute some SQL queries.
  268. # This must be in the bottom of the [beewax] section.
  269. [[query_servers]]
  270. [[[default]]]
  271. # Host where the Query Server Thrift daemon is running
  272. ## server_host=localhost
  273. # Configure the port the Query Server Thrift server
  274. ## server_port=8002
  275. # If DDL queries are supported (e.g. DROP can be sent directly to this server)
  276. ## support_ddl=True
  277. ###########################################################################
  278. # Settings to configure Job Designer
  279. ###########################################################################
  280. [jobsub]
  281. # Location on HDFS where the jobsub examples and templates are stored.
  282. ## remote_data_dir=/user/hue/jobsub
  283. # Location on local FS where examples and template are stored.
  284. ## local_data_dir=..../data
  285. # Location on local FS where sample data is stored
  286. ## sample_data_dir=...thirdparty/sample_data
  287. ###########################################################################
  288. # Settings to configure Job Browser.
  289. ###########################################################################
  290. [jobbrowser]
  291. # Share submitted jobs information with all users. If set to false,
  292. # submitted jobs are visible only to the owner and administrators.
  293. ## share_jobs=true
  294. ###########################################################################
  295. # Settings to configure the Shell application
  296. ###########################################################################
  297. [shell]
  298. # The shell_buffer_amount specifies the number of bytes of output per shell
  299. # that the Shell app will keep in memory. If not specified, it defaults to
  300. # 524288 (512 MiB).
  301. ## shell_buffer_amount=100
  302. # If you run Hue against a Hadoop cluster with Kerberos security enabled, the
  303. # Shell app needs to acquire delegation tokens for the subprocesses to work
  304. # correctly. These delegation tokens are stored as temporary files in some
  305. # directory. You can configure this directory here. If not specified, it
  306. # defaults to /tmp/hue_delegation_tokens.
  307. ## shell_delegation_token_dir=/tmp/hue_delegation_tokens
  308. [[ shelltypes ]]
  309. # Define and configure a new shell type "flume"
  310. # ------------------------------------------------------------------------
  311. [[[ flume ]]]
  312. nice_name = "Flume Shell"
  313. command = "/usr/bin/flume shell"
  314. help = "The command-line Flume client interface."
  315. [[[[ environment ]]]]
  316. # You can specify environment variables for the Flume shell
  317. # in this section.
  318. # Define and configure a new shell type "pig"
  319. # ------------------------------------------------------------------------
  320. [[[ pig ]]]
  321. nice_name = "Pig Shell (Grunt)"
  322. command = "/usr/bin/pig -l /dev/null"
  323. help = "The command-line interpreter for Pig"
  324. [[[[ environment ]]]]
  325. # You can specify environment variables for the Pig shell
  326. # in this section. Note that JAVA_HOME must be configured
  327. # for the Pig shell to run.
  328. [[[[[ JAVA_HOME ]]]]]
  329. value = "/usr/lib/jvm/java-6-sun"
  330. # Define and configure a new shell type "hbase"
  331. # ------------------------------------------------------------------------
  332. [[[ hbase ]]]
  333. nice_name = "HBase Shell"
  334. command = "/usr/bin/hbase shell"
  335. help = "The command-line HBase client interface."
  336. [[[[ environment ]]]]
  337. # You can configure environment variables for the HBase shell
  338. # in this section.
  339. ###########################################################################
  340. # Settings for the User Admin application
  341. ###########################################################################
  342. [useradmin]
  343. # The name of the default user group that users will be a member of
  344. ## default_user_group=default