|
@@ -3,56 +3,56 @@
|
|
|
|
|
|
|
|
[desktop]
|
|
[desktop]
|
|
|
|
|
|
|
|
- # Set this to a random string, the longer the better.
|
|
|
|
|
- secret_key=kasdlfjknasdfl3hbaksk3bwkasdfkasdfba23asdf
|
|
|
|
|
-
|
|
|
|
|
- # Webserver listens on this address and port
|
|
|
|
|
- http_host=0.0.0.0
|
|
|
|
|
- http_port=8888
|
|
|
|
|
-
|
|
|
|
|
- # Time zone name
|
|
|
|
|
- time_zone=America/Los_Angeles
|
|
|
|
|
-
|
|
|
|
|
- # Enable or disable debug mode.
|
|
|
|
|
- django_debug_mode=false
|
|
|
|
|
-
|
|
|
|
|
- # Enable or disable backtrace for server error
|
|
|
|
|
- http_500_debug_mode=false
|
|
|
|
|
-
|
|
|
|
|
- app_blacklist=search,hbase,security
|
|
|
|
|
-
|
|
|
|
|
- # Use gunicorn or not
|
|
|
|
|
- use_cherrypy_server=false
|
|
|
|
|
-
|
|
|
|
|
- # Gunicorn work class: gevent or evenlet, gthread or sync.
|
|
|
|
|
- gunicorn_work_class=sync
|
|
|
|
|
- gunicorn_number_of_workers=1
|
|
|
|
|
-
|
|
|
|
|
- # Configuration options for specifying the Desktop Database. For more info,
|
|
|
|
|
- # see http://docs.djangoproject.com/en/1.11/ref/settings/#database-engine
|
|
|
|
|
- # ------------------------------------------------------------------------
|
|
|
|
|
- [[database]]
|
|
|
|
|
- # Database engine is typically one of:
|
|
|
|
|
- # postgresql_psycopg2, mysql, sqlite3 or oracle.
|
|
|
|
|
- #
|
|
|
|
|
- # Note that for sqlite3, 'name', below is a path to the filename. For other backends, it is the database name
|
|
|
|
|
- # Note for Oracle, options={"threaded":true} must be set in order to avoid crashes.
|
|
|
|
|
- # Note for Oracle, you can use the Oracle Service Name by setting "host=" and "port=" and then "name=<host>:<port>/<service_name>".
|
|
|
|
|
- # Note for MariaDB use the 'mysql' engine.
|
|
|
|
|
-
|
|
|
|
|
- # engine=postgresql_psycopg2
|
|
|
|
|
- # host=hue-postgres
|
|
|
|
|
- # port=5432
|
|
|
|
|
- # user=hue
|
|
|
|
|
- # password=hue
|
|
|
|
|
- # name=hue
|
|
|
|
|
-
|
|
|
|
|
- # engine=mysql
|
|
|
|
|
- # host=database
|
|
|
|
|
- # port=3306
|
|
|
|
|
- # user=root
|
|
|
|
|
- # password=secret
|
|
|
|
|
- # name=hue
|
|
|
|
|
|
|
+# Set this to a random string, the longer the better.
|
|
|
|
|
+secret_key=kasdlfjknasdfl3hbaksk3bwkasdfkasdfba23asdf
|
|
|
|
|
+
|
|
|
|
|
+# Webserver listens on this address and port
|
|
|
|
|
+http_host=0.0.0.0
|
|
|
|
|
+http_port=8888
|
|
|
|
|
+
|
|
|
|
|
+# Time zone name
|
|
|
|
|
+time_zone=America/Los_Angeles
|
|
|
|
|
+
|
|
|
|
|
+# Enable or disable debug mode.
|
|
|
|
|
+django_debug_mode=false
|
|
|
|
|
+
|
|
|
|
|
+# Enable or disable backtrace for server error
|
|
|
|
|
+http_500_debug_mode=false
|
|
|
|
|
+
|
|
|
|
|
+app_blacklist=search,hbase,security
|
|
|
|
|
+
|
|
|
|
|
+# Use gunicorn or not
|
|
|
|
|
+use_cherrypy_server=false
|
|
|
|
|
+
|
|
|
|
|
+# Gunicorn work class: gevent or evenlet, gthread or sync.
|
|
|
|
|
+gunicorn_work_class=sync
|
|
|
|
|
+gunicorn_number_of_workers=1
|
|
|
|
|
+
|
|
|
|
|
+# Configuration options for specifying the Desktop Database. For more info,
|
|
|
|
|
+# see http://docs.djangoproject.com/en/1.11/ref/settings/#database-engine
|
|
|
|
|
+# ------------------------------------------------------------------------
|
|
|
|
|
+[[database]]
|
|
|
|
|
+# Database engine is typically one of:
|
|
|
|
|
+# postgresql_psycopg2, mysql, sqlite3 or oracle.
|
|
|
|
|
+#
|
|
|
|
|
+# Note that for sqlite3, 'name', below is a path to the filename. For other backends, it is the database name
|
|
|
|
|
+# Note for Oracle, options={"threaded":true} must be set in order to avoid crashes.
|
|
|
|
|
+# Note for Oracle, you can use the Oracle Service Name by setting "host=" and "port=" and then "name=<host>:<port>/<service_name>".
|
|
|
|
|
+# Note for MariaDB use the 'mysql' engine.
|
|
|
|
|
+
|
|
|
|
|
+# engine=postgresql_psycopg2
|
|
|
|
|
+# host=hue-postgres
|
|
|
|
|
+# port=5432
|
|
|
|
|
+# user=hue
|
|
|
|
|
+# password=hue
|
|
|
|
|
+# name=hue
|
|
|
|
|
+
|
|
|
|
|
+# engine=mysql
|
|
|
|
|
+# host=database
|
|
|
|
|
+# port=3306
|
|
|
|
|
+# user=root
|
|
|
|
|
+# password=secret
|
|
|
|
|
+# name=hue
|
|
|
|
|
|
|
|
###########################################################################
|
|
###########################################################################
|
|
|
# Settings to configure the snippets available in the Notebook
|
|
# Settings to configure the snippets available in the Notebook
|
|
@@ -60,165 +60,167 @@
|
|
|
|
|
|
|
|
# [notebook]
|
|
# [notebook]
|
|
|
|
|
|
|
|
- # One entry for each type of snippet.
|
|
|
|
|
- # [[interpreters]]
|
|
|
|
|
- # Define the name and how to connect and execute the language.
|
|
|
|
|
- # https://docs.gethue.com/administrator/configuration/editor/
|
|
|
|
|
-
|
|
|
|
|
- # Example for Docker compose
|
|
|
|
|
- # [[[mysql]]]
|
|
|
|
|
- # name = MySQL
|
|
|
|
|
- # interface=sqlalchemy
|
|
|
|
|
- # ## https://docs.sqlalchemy.org/en/latest/dialects/mysql.html
|
|
|
|
|
- # options='{"url": "mysql://root:secret@database:3306/hue"}'
|
|
|
|
|
- # ## options='{"url": "mysql://${USER}:${PASSWORD}@localhost:3306/hue"}'
|
|
|
|
|
-
|
|
|
|
|
- # [[[hive]]]
|
|
|
|
|
- # name=Hive
|
|
|
|
|
- # interface=hiveserver2
|
|
|
|
|
-
|
|
|
|
|
- # [[[impala]]]
|
|
|
|
|
- # name=Impala
|
|
|
|
|
- # interface=hiveserver2
|
|
|
|
|
-
|
|
|
|
|
- # [[[sql]]]
|
|
|
|
|
- # name=SparkSql
|
|
|
|
|
- # interface=livy
|
|
|
|
|
-
|
|
|
|
|
- # [[[spark]]]
|
|
|
|
|
- # name=Scala
|
|
|
|
|
- # interface=livy
|
|
|
|
|
-
|
|
|
|
|
- # [[[pyspark]]]
|
|
|
|
|
- # name=PySpark
|
|
|
|
|
- # interface=livy
|
|
|
|
|
-
|
|
|
|
|
- # [[[r]]]
|
|
|
|
|
- # name=R
|
|
|
|
|
- # interface=livy
|
|
|
|
|
-
|
|
|
|
|
- # [[jar]]]
|
|
|
|
|
- # name=Spark Submit Jar
|
|
|
|
|
- # interface=livy-batch
|
|
|
|
|
-
|
|
|
|
|
- # [[[py]]]
|
|
|
|
|
- # name=Spark Submit Python
|
|
|
|
|
- # interface=livy-batch
|
|
|
|
|
-
|
|
|
|
|
- # [[[text]]]
|
|
|
|
|
- # name=Text
|
|
|
|
|
- # interface=text
|
|
|
|
|
-
|
|
|
|
|
- # [[[markdown]]]
|
|
|
|
|
- # name=Markdown
|
|
|
|
|
- # interface=text
|
|
|
|
|
-
|
|
|
|
|
- # [[[sqlite]]]
|
|
|
|
|
- # name = SQLite
|
|
|
|
|
- # interface=rdbms
|
|
|
|
|
-
|
|
|
|
|
- # [[[postgresql]]]
|
|
|
|
|
- # name = PostgreSQL
|
|
|
|
|
- # interface=rdbms
|
|
|
|
|
-
|
|
|
|
|
- # [[[oracle]]]
|
|
|
|
|
- # name = Oracle
|
|
|
|
|
- # interface=rdbms
|
|
|
|
|
-
|
|
|
|
|
- # [[[solr]]]
|
|
|
|
|
- # name = Solr SQL
|
|
|
|
|
- # interface=solr
|
|
|
|
|
- # ## Name of the collection handler
|
|
|
|
|
- # # options='{"collection": "default"}'
|
|
|
|
|
-
|
|
|
|
|
- # [[[pig]]]
|
|
|
|
|
- # name=Pig
|
|
|
|
|
- # interface=oozie
|
|
|
|
|
-
|
|
|
|
|
- # [[[java]]]
|
|
|
|
|
- # name=Java
|
|
|
|
|
- # interface=oozie
|
|
|
|
|
-
|
|
|
|
|
- # [[[spark2]]]
|
|
|
|
|
- # name=Spark
|
|
|
|
|
- # interface=oozie
|
|
|
|
|
-
|
|
|
|
|
- # [[[mapreduce]]]
|
|
|
|
|
- # name=MapReduce
|
|
|
|
|
- # interface=oozie
|
|
|
|
|
-
|
|
|
|
|
- # [[[sqoop1]]]
|
|
|
|
|
- # name=Sqoop1
|
|
|
|
|
- # interface=oozie
|
|
|
|
|
-
|
|
|
|
|
- # [[[distcp]]]
|
|
|
|
|
- # name=Distcp
|
|
|
|
|
- # interface=oozie
|
|
|
|
|
-
|
|
|
|
|
- # [[[shell]]]
|
|
|
|
|
- # name=Shell
|
|
|
|
|
- # interface=oozie
|
|
|
|
|
-
|
|
|
|
|
- # [[[presto]]]
|
|
|
|
|
- # name=Presto SQL
|
|
|
|
|
- # interface=presto
|
|
|
|
|
- # ## Specific options for connecting to the Presto server.
|
|
|
|
|
- # ## The JDBC driver presto-jdbc.jar need to be in the CLASSPATH environment variable.
|
|
|
|
|
- # ## If 'user' and 'password' are omitted, they will be prompted in the UI.
|
|
|
|
|
- # options='{"url": "jdbc:presto://localhost:8080/catalog/schema", "driver": "io.prestosql.jdbc.PrestoDriver", "user": "root", "password": "root"}'
|
|
|
|
|
-
|
|
|
|
|
- # [[[clickhouse]]]
|
|
|
|
|
- # name=ClickHouse
|
|
|
|
|
- # interface=jdbc
|
|
|
|
|
- # ## Specific options for connecting to the ClickHouse server.
|
|
|
|
|
- # ## The JDBC driver clickhouse-jdbc.jar and its related jars need to be in the CLASSPATH environment variable.
|
|
|
|
|
- # options='{"url": "jdbc:clickhouse://localhost:8123", "driver": "ru.yandex.clickhouse.ClickHouseDriver", "user": "readonly", "password": ""}'
|
|
|
|
|
|
|
+# One entry for each type of snippet.
|
|
|
|
|
+# [[interpreters]]
|
|
|
|
|
+# Define the name and how to connect and execute the language.
|
|
|
|
|
+# https://docs.gethue.com/administrator/configuration/editor/
|
|
|
|
|
+
|
|
|
|
|
+# Example for Docker compose
|
|
|
|
|
+# [[[mysql]]]
|
|
|
|
|
+# name = MySQL
|
|
|
|
|
+# interface=sqlalchemy
|
|
|
|
|
+# ## https://docs.sqlalchemy.org/en/latest/dialects/mysql.html
|
|
|
|
|
+# options='{"url": "mysql://root:secret@database:3306/hue"}'
|
|
|
|
|
+# ## options='{"url": "mysql://${USER}:${PASSWORD}@localhost:3306/hue"}'
|
|
|
|
|
+
|
|
|
|
|
+# [[[hive]]]
|
|
|
|
|
+# name=Hive
|
|
|
|
|
+# interface=hiveserver2
|
|
|
|
|
+
|
|
|
|
|
+# [[[impala]]]
|
|
|
|
|
+# name=Impala
|
|
|
|
|
+# interface=hiveserver2
|
|
|
|
|
+
|
|
|
|
|
+# [[[sql]]]
|
|
|
|
|
+# name=SparkSql
|
|
|
|
|
+# interface=livy
|
|
|
|
|
+
|
|
|
|
|
+# [[[spark]]]
|
|
|
|
|
+# name=Scala
|
|
|
|
|
+# interface=livy
|
|
|
|
|
+
|
|
|
|
|
+# [[[pyspark]]]
|
|
|
|
|
+# name=PySpark
|
|
|
|
|
+# interface=livy
|
|
|
|
|
+
|
|
|
|
|
+# [[[r]]]
|
|
|
|
|
+# name=R
|
|
|
|
|
+# interface=livy
|
|
|
|
|
+
|
|
|
|
|
+# [[jar]]]
|
|
|
|
|
+# name=Spark Submit Jar
|
|
|
|
|
+# interface=livy-batch
|
|
|
|
|
+
|
|
|
|
|
+# [[[py]]]
|
|
|
|
|
+# name=Spark Submit Python
|
|
|
|
|
+# interface=livy-batch
|
|
|
|
|
+
|
|
|
|
|
+# [[[text]]]
|
|
|
|
|
+# name=Text
|
|
|
|
|
+# interface=text
|
|
|
|
|
+
|
|
|
|
|
+# [[[markdown]]]
|
|
|
|
|
+# name=Markdown
|
|
|
|
|
+# interface=text
|
|
|
|
|
+
|
|
|
|
|
+# [[[sqlite]]]
|
|
|
|
|
+# name = SQLite
|
|
|
|
|
+# interface=rdbms
|
|
|
|
|
+
|
|
|
|
|
+# [[[postgresql]]]
|
|
|
|
|
+# name = PostgreSQL
|
|
|
|
|
+# interface=rdbms
|
|
|
|
|
+
|
|
|
|
|
+# [[[oracle]]]
|
|
|
|
|
+# name = Oracle
|
|
|
|
|
+# interface=rdbms
|
|
|
|
|
+
|
|
|
|
|
+# [[[solr]]]
|
|
|
|
|
+# name = Solr SQL
|
|
|
|
|
+# interface=solr
|
|
|
|
|
+# ## Name of the collection handler
|
|
|
|
|
+# # options='{"collection": "default"}'
|
|
|
|
|
+
|
|
|
|
|
+# [[[pig]]]
|
|
|
|
|
+# name=Pig
|
|
|
|
|
+# interface=oozie
|
|
|
|
|
+
|
|
|
|
|
+# [[[java]]]
|
|
|
|
|
+# name=Java
|
|
|
|
|
+# interface=oozie
|
|
|
|
|
+
|
|
|
|
|
+# [[[spark2]]]
|
|
|
|
|
+# name=Spark
|
|
|
|
|
+# interface=oozie
|
|
|
|
|
+
|
|
|
|
|
+# [[[mapreduce]]]
|
|
|
|
|
+# name=MapReduce
|
|
|
|
|
+# interface=oozie
|
|
|
|
|
+
|
|
|
|
|
+# [[[sqoop1]]]
|
|
|
|
|
+# name=Sqoop1
|
|
|
|
|
+# interface=oozie
|
|
|
|
|
+
|
|
|
|
|
+# [[[distcp]]]
|
|
|
|
|
+# name=Distcp
|
|
|
|
|
+# interface=oozie
|
|
|
|
|
+
|
|
|
|
|
+# [[[shell]]]
|
|
|
|
|
+# name=Shell
|
|
|
|
|
+# interface=oozie
|
|
|
|
|
+
|
|
|
|
|
+# [[[presto]]]
|
|
|
|
|
+# name=Presto SQL
|
|
|
|
|
+# interface=presto
|
|
|
|
|
+# ## Specific options for connecting to the Presto server.
|
|
|
|
|
+# ## The JDBC driver presto-jdbc.jar need to be in the CLASSPATH environment variable.
|
|
|
|
|
+# ## If 'user' and 'password' are omitted, they will be prompted in the UI.
|
|
|
|
|
+# options='{"url": "jdbc:presto://localhost:8080/catalog/schema", "driver": "io.prestosql.jdbc.PrestoDriver", "user": "root", "password": "root"}'
|
|
|
|
|
+
|
|
|
|
|
+# [[[clickhouse]]]
|
|
|
|
|
+# name=ClickHouse
|
|
|
|
|
+# interface=jdbc
|
|
|
|
|
+# ## Specific options for connecting to the ClickHouse server.
|
|
|
|
|
+# ## The JDBC driver clickhouse-jdbc.jar and its related jars need to be in the CLASSPATH environment variable.
|
|
|
|
|
+# options='{"url": "jdbc:clickhouse://localhost:8123", "driver": "ru.yandex.clickhouse.ClickHouseDriver", "user": "readonly", "password": ""}'
|
|
|
|
|
|
|
|
|
|
|
|
|
[dashboard]
|
|
[dashboard]
|
|
|
|
|
|
|
|
- # Activate the SQL Dashboard (beta).
|
|
|
|
|
- has_sql_enabled=true
|
|
|
|
|
|
|
+# Activate the SQL Dashboard (beta).
|
|
|
|
|
+has_sql_enabled=true
|
|
|
|
|
|
|
|
|
|
|
|
|
[hadoop]
|
|
[hadoop]
|
|
|
|
|
|
|
|
- # Configuration for HDFS NameNode
|
|
|
|
|
- # ------------------------------------------------------------------------
|
|
|
|
|
- [[hdfs_clusters]]
|
|
|
|
|
- # HA support by using HttpFs
|
|
|
|
|
|
|
+# Configuration for HDFS NameNode
|
|
|
|
|
+# ------------------------------------------------------------------------
|
|
|
|
|
+[[hdfs_clusters]]
|
|
|
|
|
+# HA support by using HttpFs
|
|
|
|
|
|
|
|
- # [[[default]]]
|
|
|
|
|
- # Enter the filesystem uri
|
|
|
|
|
- # fs_defaultfs=hdfs://localhost:8020
|
|
|
|
|
|
|
+# [[[default]]]
|
|
|
|
|
+# Enter the filesystem uri
|
|
|
|
|
+# fs_defaultfs=hdfs://localhost:8020
|
|
|
|
|
|
|
|
- # Use WebHdfs/HttpFs as the communication mechanism.
|
|
|
|
|
- # Domain should be the NameNode or HttpFs host.
|
|
|
|
|
- # Default port is 14000 for HttpFs.
|
|
|
|
|
- ## webhdfs_url=http://localhost:50070/webhdfs/v1
|
|
|
|
|
|
|
+# Use WebHdfs/HttpFs as the communication mechanism.
|
|
|
|
|
+# Domain should be the NameNode or HttpFs host.
|
|
|
|
|
+# Default port is 14000 for HttpFs.
|
|
|
|
|
+## webhdfs_url=http://localhost:50070/webhdfs/v1
|
|
|
|
|
|
|
|
- # Configuration for YARN (MR2)
|
|
|
|
|
- # ------------------------------------------------------------------------
|
|
|
|
|
- [[yarn_clusters]]
|
|
|
|
|
|
|
+is_enabled=false
|
|
|
|
|
|
|
|
- # [[[default]]]
|
|
|
|
|
- # Enter the host on which you are running the ResourceManager
|
|
|
|
|
- ## resourcemanager_host=localhost
|
|
|
|
|
|
|
+# Configuration for YARN (MR2)
|
|
|
|
|
+# ------------------------------------------------------------------------
|
|
|
|
|
+[[yarn_clusters]]
|
|
|
|
|
|
|
|
- # The port where the ResourceManager IPC listens on
|
|
|
|
|
- ## resourcemanager_port=8032
|
|
|
|
|
|
|
+# [[[default]]]
|
|
|
|
|
+# Enter the host on which you are running the ResourceManager
|
|
|
|
|
+## resourcemanager_host=localhost
|
|
|
|
|
|
|
|
- # URL of the ResourceManager API
|
|
|
|
|
- ## resourcemanager_api_url=http://localhost:8088
|
|
|
|
|
|
|
+# The port where the ResourceManager IPC listens on
|
|
|
|
|
+## resourcemanager_port=8032
|
|
|
|
|
|
|
|
- # URL of the ProxyServer API
|
|
|
|
|
- ## proxy_api_url=http://localhost:8088
|
|
|
|
|
|
|
+# URL of the ResourceManager API
|
|
|
|
|
+## resourcemanager_api_url=http://localhost:8088
|
|
|
|
|
|
|
|
- # URL of the HistoryServer API
|
|
|
|
|
- ## history_server_api_url=http://localhost:19888
|
|
|
|
|
|
|
+# URL of the ProxyServer API
|
|
|
|
|
+## proxy_api_url=http://localhost:8088
|
|
|
|
|
|
|
|
- # URL of the Spark History Server
|
|
|
|
|
- ## spark_history_server_url=http://localhost:18088
|
|
|
|
|
|
|
+# URL of the HistoryServer API
|
|
|
|
|
+## history_server_api_url=http://localhost:19888
|
|
|
|
|
+
|
|
|
|
|
+# URL of the Spark History Server
|
|
|
|
|
+## spark_history_server_url=http://localhost:18088
|
|
|
|
|
|
|
|
|
|
|
|
|
###########################################################################
|
|
###########################################################################
|
|
@@ -227,12 +229,12 @@
|
|
|
|
|
|
|
|
[beeswax]
|
|
[beeswax]
|
|
|
|
|
|
|
|
- # Host where HiveServer2 is running.
|
|
|
|
|
- # If Kerberos security is enabled, use fully-qualified domain name (FQDN).
|
|
|
|
|
- ## hive_server_host=localhost
|
|
|
|
|
|
|
+# Host where HiveServer2 is running.
|
|
|
|
|
+# If Kerberos security is enabled, use fully-qualified domain name (FQDN).
|
|
|
|
|
+## hive_server_host=localhost
|
|
|
|
|
|
|
|
- # Port where HiveServer2 Thrift server runs on.
|
|
|
|
|
- ## hive_server_port=10000
|
|
|
|
|
|
|
+# Port where HiveServer2 Thrift server runs on.
|
|
|
|
|
+## hive_server_port=10000
|
|
|
|
|
|
|
|
|
|
|
|
|
###########################################################################
|
|
###########################################################################
|
|
@@ -240,11 +242,11 @@
|
|
|
###########################################################################
|
|
###########################################################################
|
|
|
|
|
|
|
|
[impala]
|
|
[impala]
|
|
|
- # Host of the Impala Server (one of the Impalad)
|
|
|
|
|
- ## server_host=localhost
|
|
|
|
|
|
|
+# Host of the Impala Server (one of the Impalad)
|
|
|
|
|
+## server_host=localhost
|
|
|
|
|
|
|
|
- # Port of the Impala Server
|
|
|
|
|
- ## server_port=21050
|
|
|
|
|
|
|
+# Port of the Impala Server
|
|
|
|
|
+## server_port=21050
|
|
|
|
|
|
|
|
|
|
|
|
|
###########################################################################
|
|
###########################################################################
|
|
@@ -252,23 +254,23 @@
|
|
|
###########################################################################
|
|
###########################################################################
|
|
|
|
|
|
|
|
[spark]
|
|
[spark]
|
|
|
- # The Livy Server URL.
|
|
|
|
|
- ## livy_server_url=http://localhost:8998
|
|
|
|
|
|
|
+# The Livy Server URL.
|
|
|
|
|
+## livy_server_url=http://localhost:8998
|
|
|
|
|
|
|
|
- # Configure Livy to start in local 'process' mode, or 'yarn' workers.
|
|
|
|
|
- ## livy_server_session_kind=yarn
|
|
|
|
|
|
|
+# Configure Livy to start in local 'process' mode, or 'yarn' workers.
|
|
|
|
|
+## livy_server_session_kind=yarn
|
|
|
|
|
|
|
|
- # Whether Livy requires client to perform Kerberos authentication.
|
|
|
|
|
- ## security_enabled=false
|
|
|
|
|
|
|
+# Whether Livy requires client to perform Kerberos authentication.
|
|
|
|
|
+## security_enabled=false
|
|
|
|
|
|
|
|
- # Host of the Sql Server
|
|
|
|
|
- ## sql_server_host=localhost
|
|
|
|
|
|
|
+# Host of the Sql Server
|
|
|
|
|
+## sql_server_host=localhost
|
|
|
|
|
|
|
|
- # Port of the Sql Server
|
|
|
|
|
- ## sql_server_port=10000
|
|
|
|
|
|
|
+# Port of the Sql Server
|
|
|
|
|
+## sql_server_port=10000
|
|
|
|
|
|
|
|
- # Choose whether Hue should validate certificates received from the server.
|
|
|
|
|
- ## ssl_cert_ca_verify=true
|
|
|
|
|
|
|
+# Choose whether Hue should validate certificates received from the server.
|
|
|
|
|
+## ssl_cert_ca_verify=true
|
|
|
|
|
|
|
|
|
|
|
|
|
###########################################################################
|
|
###########################################################################
|
|
@@ -276,8 +278,8 @@
|
|
|
###########################################################################
|
|
###########################################################################
|
|
|
|
|
|
|
|
[hbase]
|
|
[hbase]
|
|
|
- # Comma-separated list of HBase Thrift servers for clusters in the format of '(name|host:port)'.
|
|
|
|
|
- ## hbase_clusters=(Cluster|localhost:9090)
|
|
|
|
|
|
|
+# Comma-separated list of HBase Thrift servers for clusters in the format of '(name|host:port)'.
|
|
|
|
|
+## hbase_clusters=(Cluster|localhost:9090)
|
|
|
|
|
|
|
|
|
|
|
|
|
###########################################################################
|
|
###########################################################################
|
|
@@ -286,8 +288,8 @@
|
|
|
|
|
|
|
|
[search]
|
|
[search]
|
|
|
|
|
|
|
|
- # URL of the Solr Server
|
|
|
|
|
- ## solr_url=http://localhost:8983/solr/
|
|
|
|
|
|
|
+# URL of the Solr Server
|
|
|
|
|
+## solr_url=http://localhost:8983/solr/
|
|
|
|
|
|
|
|
|
|
|
|
|
###########################################################################
|
|
###########################################################################
|
|
@@ -295,9 +297,9 @@
|
|
|
###########################################################################
|
|
###########################################################################
|
|
|
|
|
|
|
|
[liboozie]
|
|
[liboozie]
|
|
|
- # The URL where the Oozie service runs on. This is required in order for
|
|
|
|
|
- # users to submit jobs. Empty value disables the config check.
|
|
|
|
|
- ## oozie_url=http://localhost:11000/oozie
|
|
|
|
|
|
|
+# The URL where the Oozie service runs on. This is required in order for
|
|
|
|
|
+# users to submit jobs. Empty value disables the config check.
|
|
|
|
|
+## oozie_url=http://localhost:11000/oozie
|
|
|
|
|
|
|
|
|
|
|
|
|
###########################################################################
|
|
###########################################################################
|
|
@@ -305,54 +307,54 @@
|
|
|
###########################################################################
|
|
###########################################################################
|
|
|
|
|
|
|
|
[aws]
|
|
[aws]
|
|
|
- [[aws_accounts]]
|
|
|
|
|
- # Default AWS account
|
|
|
|
|
- ## [[[default]]]
|
|
|
|
|
- # AWS credentials
|
|
|
|
|
- ## access_key_id=
|
|
|
|
|
- ## secret_access_key=
|
|
|
|
|
- ## security_token=
|
|
|
|
|
|
|
+[[aws_accounts]]
|
|
|
|
|
+# Default AWS account
|
|
|
|
|
+## [[[default]]]
|
|
|
|
|
+# AWS credentials
|
|
|
|
|
+## access_key_id=
|
|
|
|
|
+## secret_access_key=
|
|
|
|
|
+## security_token=
|
|
|
|
|
|
|
|
- # Execute this script to produce the AWS access key ID.
|
|
|
|
|
- ## access_key_id_script=/path/access_key_id.sh
|
|
|
|
|
|
|
+# Execute this script to produce the AWS access key ID.
|
|
|
|
|
+## access_key_id_script=/path/access_key_id.sh
|
|
|
|
|
|
|
|
- # Execute this script to produce the AWS secret access key.
|
|
|
|
|
- ## secret_access_key_script=/path/secret_access_key.sh
|
|
|
|
|
|
|
+# Execute this script to produce the AWS secret access key.
|
|
|
|
|
+## secret_access_key_script=/path/secret_access_key.sh
|
|
|
|
|
|
|
|
- # Allow to use either environment variables or
|
|
|
|
|
- # EC2 InstanceProfile to retrieve AWS credentials.
|
|
|
|
|
- ## allow_environment_credentials=yes
|
|
|
|
|
|
|
+# Allow to use either environment variables or
|
|
|
|
|
+# EC2 InstanceProfile to retrieve AWS credentials.
|
|
|
|
|
+## allow_environment_credentials=yes
|
|
|
|
|
|
|
|
- # AWS region to use, if no region is specified, will attempt to connect to standard s3.amazonaws.com endpoint
|
|
|
|
|
- ## region=us-east-1
|
|
|
|
|
|
|
+# AWS region to use, if no region is specified, will attempt to connect to standard s3.amazonaws.com endpoint
|
|
|
|
|
+## region=us-east-1
|
|
|
|
|
|
|
|
- # Endpoint overrides
|
|
|
|
|
- ## host=
|
|
|
|
|
|
|
+# Endpoint overrides
|
|
|
|
|
+## host=
|
|
|
|
|
|
|
|
- # Proxy address and port
|
|
|
|
|
- ## proxy_address=
|
|
|
|
|
- ## proxy_port=8080
|
|
|
|
|
- ## proxy_user=
|
|
|
|
|
- ## proxy_pass=
|
|
|
|
|
|
|
+# Proxy address and port
|
|
|
|
|
+## proxy_address=
|
|
|
|
|
+## proxy_port=8080
|
|
|
|
|
+## proxy_user=
|
|
|
|
|
+## proxy_pass=
|
|
|
|
|
|
|
|
- # Secure connections are the default, but this can be explicitly overridden:
|
|
|
|
|
- ## is_secure=true
|
|
|
|
|
|
|
+# Secure connections are the default, but this can be explicitly overridden:
|
|
|
|
|
+## is_secure=true
|
|
|
|
|
|
|
|
|
|
|
|
|
###########################################################################
|
|
###########################################################################
|
|
|
# Settings for the Azure lib
|
|
# Settings for the Azure lib
|
|
|
###########################################################################
|
|
###########################################################################
|
|
|
[azure]
|
|
[azure]
|
|
|
- [[azure_accounts]]
|
|
|
|
|
- # [[[default]]]
|
|
|
|
|
- # client_id=xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxx
|
|
|
|
|
- # client_secret=xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx
|
|
|
|
|
- # tenant_id=xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxx
|
|
|
|
|
|
|
+[[azure_accounts]]
|
|
|
|
|
+# [[[default]]]
|
|
|
|
|
+# client_id=xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxx
|
|
|
|
|
+# client_secret=xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx
|
|
|
|
|
+# tenant_id=xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxx
|
|
|
|
|
|
|
|
- # [[abfs_clusters]]
|
|
|
|
|
- # [[[default]]]
|
|
|
|
|
- # fs_defaultfs=abfs://account@account.dfs.core.windows.net
|
|
|
|
|
- # webhdfs_url=https://account.dfs.core.windows.net/
|
|
|
|
|
|
|
+# [[abfs_clusters]]
|
|
|
|
|
+# [[[default]]]
|
|
|
|
|
+# fs_defaultfs=abfs://account@account.dfs.core.windows.net
|
|
|
|
|
+# webhdfs_url=https://account.dfs.core.windows.net/
|
|
|
|
|
|
|
|
|
|
|
|
|
###########################################################################
|
|
###########################################################################
|
|
@@ -361,22 +363,22 @@
|
|
|
|
|
|
|
|
[metadata]
|
|
[metadata]
|
|
|
|
|
|
|
|
- [[navigator]]
|
|
|
|
|
- # Navigator API URL (without version suffix).
|
|
|
|
|
- ## api_url=http://localhost:7187/api
|
|
|
|
|
-
|
|
|
|
|
- # Which authentication to use: CM or external via LDAP or SAML.
|
|
|
|
|
- ## navmetadataserver_auth_type=CMDB
|
|
|
|
|
-
|
|
|
|
|
- # Username of the CM user used for authentication.
|
|
|
|
|
- ## navmetadataserver_cmdb_user=hue
|
|
|
|
|
- # CM password of the user used for authentication.
|
|
|
|
|
- ## navmetadataserver_cmdb_password=
|
|
|
|
|
- # Execute this script to produce the CM password. This will be used when the plain password is not set.
|
|
|
|
|
- # navmetadataserver_cmdb_password_script=
|
|
|
|
|
-
|
|
|
|
|
- # [[atlas]]
|
|
|
|
|
- # interface=atlas
|
|
|
|
|
- # api_url=http://localhost:21000/api/atlas/v2
|
|
|
|
|
- # server_user=admin
|
|
|
|
|
- # server_password=admin
|
|
|
|
|
|
|
+[[navigator]]
|
|
|
|
|
+# Navigator API URL (without version suffix).
|
|
|
|
|
+## api_url=http://localhost:7187/api
|
|
|
|
|
+
|
|
|
|
|
+# Which authentication to use: CM or external via LDAP or SAML.
|
|
|
|
|
+## navmetadataserver_auth_type=CMDB
|
|
|
|
|
+
|
|
|
|
|
+# Username of the CM user used for authentication.
|
|
|
|
|
+## navmetadataserver_cmdb_user=hue
|
|
|
|
|
+# CM password of the user used for authentication.
|
|
|
|
|
+## navmetadataserver_cmdb_password=
|
|
|
|
|
+# Execute this script to produce the CM password. This will be used when the plain password is not set.
|
|
|
|
|
+# navmetadataserver_cmdb_password_script=
|
|
|
|
|
+
|
|
|
|
|
+# [[atlas]]
|
|
|
|
|
+# interface=atlas
|
|
|
|
|
+# api_url=http://localhost:21000/api/atlas/v2
|
|
|
|
|
+# server_user=admin
|
|
|
|
|
+# server_password=admin
|