z-hue-overrides.ini 10.0 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391
  1. # Lightweight Hue configuration file
  2. # ==================================
  3. [desktop]
  4. # Set this to a random string, the longer the better.
  5. secret_key=kasdlfjknasdfl3hbaksk3bwkasdfkasdfba23asdf
  6. # Webserver listens on this address and port
  7. http_host=0.0.0.0
  8. http_port=8888
  9. # Time zone name
  10. time_zone=America/Los_Angeles
  11. # Enable or disable debug mode.
  12. django_debug_mode=false
  13. # Enable or disable backtrace for server error
  14. http_500_debug_mode=false
  15. app_blacklist=search,hbase,security
  16. # Use gunicorn or not
  17. use_cherrypy_server=false
  18. # Gunicorn work class: gevent or evenlet, gthread or sync.
  19. gunicorn_work_class=sync
  20. gunicorn_number_of_workers=1
  21. # Configuration options for specifying the Desktop Database. For more info,
  22. # see http://docs.djangoproject.com/en/1.11/ref/settings/#database-engine
  23. # ------------------------------------------------------------------------
  24. [[database]]
  25. # Database engine is typically one of:
  26. # postgresql_psycopg2, mysql, sqlite3 or oracle.
  27. #
  28. # Note that for sqlite3, 'name', below is a path to the filename. For other backends, it is the database name
  29. # Note for Oracle, options={"threaded":true} must be set in order to avoid crashes.
  30. # Note for Oracle, you can use the Oracle Service Name by setting "host=" and "port=" and then "name=<host>:<port>/<service_name>".
  31. # Note for MariaDB use the 'mysql' engine.
  32. # engine=postgresql_psycopg2
  33. # host=hue-postgres
  34. # port=5432
  35. # user=hue
  36. # password=hue
  37. # name=hue
  38. # engine=mysql
  39. # host=database
  40. # port=3306
  41. # user=root
  42. # password=secret
  43. # name=hue
  44. ###########################################################################
  45. # Settings to configure the snippets available in the Notebook
  46. ###########################################################################
  47. # [notebook]
  48. # One entry for each type of snippet.
  49. # [[interpreters]]
  50. # Define the name and how to connect and execute the language.
  51. # https://docs.gethue.com/administrator/configuration/editor/
  52. # Example for Docker compose
  53. # [[[mysql]]]
  54. # name = MySQL
  55. # interface=sqlalchemy
  56. # ## https://docs.sqlalchemy.org/en/latest/dialects/mysql.html
  57. # options='{"url": "mysql://root:secret@database:3306/hue"}'
  58. # ## options='{"url": "mysql://${USER}:${PASSWORD}@localhost:3306/hue"}'
  59. # [[[hive]]]
  60. # name=Hive
  61. # interface=hiveserver2
  62. # [[[impala]]]
  63. # name=Impala
  64. # interface=hiveserver2
  65. # [[[sparksql]]]
  66. # name = Spark Sql
  67. # interface=sqlalchemy
  68. # options='{"url": "hive://user:password@localhost:10000/database"}'
  69. # [[[sparksql]]]
  70. # name=SparkSql
  71. # interface=livy
  72. # [[[spark]]]
  73. # name=Scala
  74. # interface=livy
  75. # [[[pyspark]]]
  76. # name=PySpark
  77. # interface=livy
  78. # [[[r]]]
  79. # name=R
  80. # interface=livy
  81. # [[jar]]]
  82. # name=Spark Submit Jar
  83. # interface=livy-batch
  84. # [[[py]]]
  85. # name=Spark Submit Python
  86. # interface=livy-batch
  87. # [[[text]]]
  88. # name=Text
  89. # interface=text
  90. # [[[markdown]]]
  91. # name=Markdown
  92. # interface=text
  93. # [[[sqlite]]]
  94. # name = SQLite
  95. # interface=rdbms
  96. # [[[postgresql]]]
  97. # name = PostgreSQL
  98. # interface=rdbms
  99. # [[[oracle]]]
  100. # name = Oracle
  101. # interface=rdbms
  102. # [[[solr]]]
  103. # name = Solr SQL
  104. # interface=solr
  105. # ## Name of the collection handler
  106. # # options='{"collection": "default"}'
  107. # [[[pig]]]
  108. # name=Pig
  109. # interface=oozie
  110. # [[[java]]]
  111. # name=Java
  112. # interface=oozie
  113. # [[[spark2]]]
  114. # name=Spark
  115. # interface=oozie
  116. # [[[mapreduce]]]
  117. # name=MapReduce
  118. # interface=oozie
  119. # [[[sqoop1]]]
  120. # name=Sqoop1
  121. # interface=oozie
  122. # [[[distcp]]]
  123. # name=Distcp
  124. # interface=oozie
  125. # [[[shell]]]
  126. # name=Shell
  127. # interface=oozie
  128. # [[[presto]]]
  129. # name=Presto SQL
  130. # interface=presto
  131. # ## Specific options for connecting to the Presto server.
  132. # ## The JDBC driver presto-jdbc.jar need to be in the CLASSPATH environment variable.
  133. # ## If 'user' and 'password' are omitted, they will be prompted in the UI.
  134. # options='{"url": "jdbc:presto://localhost:8080/catalog/schema", "driver": "io.prestosql.jdbc.PrestoDriver", "user": "root", "password": "root"}'
  135. # [[[clickhouse]]]
  136. # name=ClickHouse
  137. # interface=jdbc
  138. # ## Specific options for connecting to the ClickHouse server.
  139. # ## The JDBC driver clickhouse-jdbc.jar and its related jars need to be in the CLASSPATH environment variable.
  140. # options='{"url": "jdbc:clickhouse://localhost:8123", "driver": "ru.yandex.clickhouse.ClickHouseDriver", "user": "readonly", "password": ""}'
  141. [dashboard]
  142. # Activate the SQL Dashboard (beta).
  143. has_sql_enabled=true
  144. [hadoop]
  145. # Configuration for HDFS NameNode
  146. # ------------------------------------------------------------------------
  147. [[hdfs_clusters]]
  148. # HA support by using HttpFs
  149. # [[[default]]]
  150. # Enter the filesystem uri
  151. ## fs_defaultfs=hdfs://localhost:8020
  152. # Use WebHdfs/HttpFs as the communication mechanism.
  153. # Domain should be the NameNode or HttpFs host.
  154. # Default port is 14000 for HttpFs.
  155. ## webhdfs_url=http://localhost:50070/webhdfs/v1
  156. ## is_enabled=false
  157. # Configuration for YARN (MR2)
  158. # ------------------------------------------------------------------------
  159. [[yarn_clusters]]
  160. # [[[default]]]
  161. # Enter the host on which you are running the ResourceManager
  162. ## resourcemanager_host=localhost
  163. # The port where the ResourceManager IPC listens on
  164. ## resourcemanager_port=8032
  165. # URL of the ResourceManager API
  166. ## resourcemanager_api_url=http://localhost:8088
  167. # URL of the ProxyServer API
  168. ## proxy_api_url=http://localhost:8088
  169. # URL of the HistoryServer API
  170. ## history_server_api_url=http://localhost:19888
  171. # URL of the Spark History Server
  172. ## spark_history_server_url=http://localhost:18088
  173. ###########################################################################
  174. # Settings to configure Beeswax with Hive
  175. ###########################################################################
  176. [beeswax]
  177. # Host where HiveServer2 is running.
  178. # If Kerberos security is enabled, use fully-qualified domain name (FQDN).
  179. ## hive_server_host=localhost
  180. # Port where HiveServer2 Thrift server runs on.
  181. ## hive_server_port=10000
  182. ###########################################################################
  183. # Settings to configure Impala
  184. ###########################################################################
  185. [impala]
  186. # Host of the Impala Server (one of the Impalad)
  187. ## server_host=localhost
  188. # Port of the Impala Server
  189. ## server_port=21050
  190. ###########################################################################
  191. # Settings to configure the Spark application.
  192. ###########################################################################
  193. [spark]
  194. # The Livy Server URL.
  195. ## livy_server_url=http://localhost:8998
  196. # Configure Livy to start in local 'process' mode, or 'yarn' workers.
  197. ## livy_server_session_kind=yarn
  198. # Whether Livy requires client to perform Kerberos authentication.
  199. ## security_enabled=false
  200. # Host of the Sql Server
  201. ## sql_server_host=localhost
  202. # Port of the Sql Server
  203. ## sql_server_port=10000
  204. # Choose whether Hue should validate certificates received from the server.
  205. ## ssl_cert_ca_verify=true
  206. ###########################################################################
  207. # Settings to configure HBase Browser
  208. ###########################################################################
  209. [hbase]
  210. # Comma-separated list of HBase Thrift servers for clusters in the format of '(name|host:port)'.
  211. ## hbase_clusters=(Cluster|localhost:9090)
  212. ###########################################################################
  213. # Settings to configure Solr Search
  214. ###########################################################################
  215. [search]
  216. # URL of the Solr Server
  217. ## solr_url=http://localhost:8983/solr/
  218. ###########################################################################
  219. # Settings to configure liboozie
  220. ###########################################################################
  221. [liboozie]
  222. # The URL where the Oozie service runs on. This is required in order for
  223. # users to submit jobs. Empty value disables the config check.
  224. ## oozie_url=http://localhost:11000/oozie
  225. ###########################################################################
  226. # Settings for the AWS lib
  227. ###########################################################################
  228. [aws]
  229. [[aws_accounts]]
  230. # Default AWS account
  231. ## [[[default]]]
  232. # AWS credentials
  233. ## access_key_id=
  234. ## secret_access_key=
  235. ## security_token=
  236. # Execute this script to produce the AWS access key ID.
  237. ## access_key_id_script=/path/access_key_id.sh
  238. # Execute this script to produce the AWS secret access key.
  239. ## secret_access_key_script=/path/secret_access_key.sh
  240. # Allow to use either environment variables or
  241. # EC2 InstanceProfile to retrieve AWS credentials.
  242. ## allow_environment_credentials=yes
  243. # AWS region to use, if no region is specified, will attempt to connect to standard s3.amazonaws.com endpoint
  244. ## region=us-east-1
  245. # Endpoint overrides
  246. ## host=
  247. # Proxy address and port
  248. ## proxy_address=
  249. ## proxy_port=8080
  250. ## proxy_user=
  251. ## proxy_pass=
  252. # Secure connections are the default, but this can be explicitly overridden:
  253. ## is_secure=true
  254. ###########################################################################
  255. # Settings for the Azure lib
  256. ###########################################################################
  257. [azure]
  258. [[azure_accounts]]
  259. # [[[default]]]
  260. # client_id=xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxx
  261. # client_secret=xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx
  262. # tenant_id=xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxx
  263. # [[abfs_clusters]]
  264. # [[[default]]]
  265. # fs_defaultfs=abfs://account@account.dfs.core.windows.net
  266. # webhdfs_url=https://account.dfs.core.windows.net/
  267. ###########################################################################
  268. # Settings to configure Metadata
  269. ###########################################################################
  270. [metadata]
  271. [[navigator]]
  272. # Navigator API URL (without version suffix).
  273. ## api_url=http://localhost:7187/api
  274. # Which authentication to use: CM or external via LDAP or SAML.
  275. ## navmetadataserver_auth_type=CMDB
  276. # Username of the CM user used for authentication.
  277. ## navmetadataserver_cmdb_user=hue
  278. # CM password of the user used for authentication.
  279. ## navmetadataserver_cmdb_password=
  280. # Execute this script to produce the CM password. This will be used when the plain password is not set.
  281. # navmetadataserver_cmdb_password_script=
  282. # [[atlas]]
  283. # interface=atlas
  284. # api_url=http://localhost:21000/api/atlas/v2
  285. # server_user=admin
  286. # server_password=admin