---
title: How to deploy Hue on HDP
author: admin
type: post
date: 2014-12-16T19:36:47+00:00
url: /how-to-deploy-hue-on-hdp/
sf_thumbnail_type:
- none
sf_thumbnail_link_type:
- link_to_post
sf_detail_type:
- none
sf_page_title:
- 1
sf_page_title_style:
- standard
sf_no_breadcrumbs:
- 1
sf_page_title_bg:
- none
sf_page_title_text_style:
- light
sf_background_image_size:
- cover
sf_social_sharing:
- 1
sf_sidebar_config:
- left-sidebar
sf_left_sidebar:
- Sidebar-2
sf_right_sidebar:
- Sidebar-1
sf_caption_position:
- caption-right
sf_remove_promo_bar:
- 1
slide_template:
- default
categories:
- Administration
---
_Guest post from Andrew that we regularly update (Dec 19th 2014)_
I decided to deploy Hue 3.7, from tarballs (note, other sources like packages from the 'Install' menu above would work too), on [HDP 2.2][1] recently and wanted to document some notes for anyone else looking to do the same.
Deployment Background:
* Node Operating System: CentOS 6.6 - 64bit
* Cluster Manager: Ambari 1.7
* Distribution: HDP 2.2
* Install Path (default): /usr/local/hue
* HUE User: hue
After compiling (some hints there), you may run into out of the box/post-compile startup issues.
* Be sure to set the appropriate Hue proxy user/groups properties in your Hadoop service configurations (e.g. WebHDFS/WebHCat/Oozie/etc)
* Don't forget to configure your Hue configuration file ('/usr/local/hue/desktop/conf/
][2]
# Startup
Hue uses an SQLite database by default and you may find the following error when attempting to connect to HUE at its default port (e.g. fqdn:8888)
* **_..._** _File "/usr/local/hue/build/env/lib/
{{< highlight bash >}}[desktop]
app_blacklist=impala
{{< /highlight >}}
With Ambari, you can review your cluster's HDFS configuration, specifically under the "Custom core-site.xml" subsection:
There should be two (2) new/custom properties added to support the HUE File Browser:
{{< highlight xml >}}
{{< /highlight >}}
With Ambari, you can go to the HDFS service settings and find this under "General"
- The property name is dfs.webhdfs.enabled ("WebHDFS enabled), and should be set to "true" by default.
- If a change is required, save the change and start/restart the service with the updated configuration.
Ensure the HDFS service is started and operating normally.
- You could quickly check some things, such as HDFS and WebHDFS by checking the WebHDFS page:
- http://
Check if the processes are running using a shell command on your NameNode:
- 'ps -ef | grep "NameNode"
{{< highlight bash >}}"Server does not support GetLog()"{{< /highlight >}}
{{< highlight bash >}}[beeswax]
# Choose whether Hue uses the GetLog() thrift call to retrieve Hive logs.
# If false, Hue will use the FetchResults() thrift call instead.
use_get_log_api=false
{{< /highlight >}}
By default, Hadoop 2.4.0 does not enable HDFS file access control lists (FACLs)
Currently not tested (should work with Thrift Server 1)
Note: when Oozie is deployed via Ambari 1.7, for HDP 2.2, the sharelib files typically found at /usr/lib/oozie/ are missing, and in turn are not staged at hdfs:/user/oozie/share/lib/ ...
{{< highlight bash >}}2014-12-15 23:32:17,626 INFO ActionStartXCommand:543 - SERVER[hdptest.construct.dev] USER[amo] GROUP[-] TOKEN[] APP[pig-app-hue-script] JOB[0000001-141215230246520-
2014-12-15 23:32:17,627 INFO ActionStartXCommand:543 - SERVER[hdptest.construct.dev] USER[amo] GROUP[-] TOKEN[] APP[pig-app-hue-script] JOB[0000001-141215230246520-
2014-12-15 23:32:17,627 INFO ActionStartXCommand:543 - SERVER[hdptest.construct.dev] USER[amo] GROUP[-] TOKEN[] APP[pig-app-hue-script] JOB[0000001-141215230246520-
2014-12-15 23:32:17,873 INFO ActionStartXCommand:543 - SERVER[hdptest.construct.dev] USER[amo] GROUP[-] TOKEN[] APP[pig-app-hue-script] JOB[0000001-141215230246520-
{{< /highlight >}}