Эх сурвалжийг харах

HUE-2890 [doc] Clean-up old SDK

Romain Rigaux 8 жил өмнө
parent
commit
0452ccee50

+ 36 - 147
docs/sdk/sdk.md

@@ -75,13 +75,8 @@ On the Front-End
 
 Hue provides a front-end framework based on
 [Bootstrap](http://twitter.github.com/bootstrap/) and
-[jQuery](http://jquery.com/).
+[Knockout js](http://knockoutjs.com/).
 
-If you are used to the Hue 1.x front-end, this is a major difference. All
-application pages are full screen requests from the browser. The HTML generated
-by your application's template is directly rendered. You do not need to
-worry about interference from another application. And you have more freedom to
-customize the front-end behavior of your application.
 
 An Architectural View
 ---------------------
@@ -96,12 +91,12 @@ may interact.
 
 The absolute minimum that you must implement (besides
 boilerplate), is a
-"Django [view](https://docs.djangoproject.com/en/1.2/topics/http/views/)"
+"Django [view](https://docs.djangoproject.com/en/1.11/#the-view-layer/)"
 function that processes the request and the associated template
 to render the response into HTML.
 
 Many apps will evolve to have a bit of custom JavaScript and
-CSS styles.  Apps that need to talk to an external service
+CSS styles. Apps that need to talk to an external service
 will pull in the code necessary to talk to that service.
 
 File Layout
@@ -146,8 +141,8 @@ Software
 --------
 
 Developing for the Hue SDK has similar requirements to running
-Hue itself.  We require python (2.6 to 2.7), Django (1.4 included
-with our distribution), Hadoop (Apache Hadoop 1.2+), Java (Sun Java 1.7),
+Hue itself.  We require python (2.6 to 2.7), Django (1.6 included
+with our distribution), Hadoop (Apache Hadoop 2+), Java (Sun Java 1.8),
 and Browser (latest Chrome, Firefox or IE9+).
 
 Recommended Reading / Important Technologies
@@ -157,7 +152,7 @@ The following are core technologies used inside of Hue.
 
 * Python.  <a href="http://diveintopython.net/">Dive Into Python</a> is one of
   several excellent books on python.
-* Django.  Start with [The Django Tutorial](http://docs.djangoproject.com/en/1.2/intro/tutorial01/).
+* Django.  Start with [The Django Tutorial](https://docs.djangoproject.com/en/1.11/intro/).
 * [Thrift](http://incubator.apache.org/thrift/) is used for communication
   between daemons.
 * [Mako](http://www.makotemplates.org/) is the preferred templating language.
@@ -172,26 +167,26 @@ Download, Unpack, Build Distro
 ------------------------------
 
 The Hue SDK is available from [Github](http://github.com/cloudera/hue). Releases
-can be found on the [download page](https://github.com/cloudera/hue/downloads).
+can be found on the [download page](http://gethue.com/category/release/).
 Releases are missing a few dependencies that could not be included because of
 licencing issues. So if you prefer to have an environment ready from scratch,
 it is preferable to checkout a particular release tag instead.
 
-    $ cd hue
+    cd hue
     ## Build
-    $ make apps
+    make apps
     ## Run
-    $ build/env/bin/hue runserver
+    build/env/bin/hue runserver
     ## Alternative run
-    $ build/env/bin/hue supervisor
+    build/env/bin/hue supervisor
     ## Visit http://localhost:8000/ with your web browser.
 
 
 Run "create_desktop_app" to Set up a New Source Tree
 --------------------------------------------
 
-    $ ./build/env/bin/hue create_desktop_app calculator
-    $ find calculator -type f
+    ./build/env/bin/hue create_desktop_app calculator
+    find calculator -type f
     calculator/setup.py                                 # distutils setup file
     calculator/src/calculator/__init__.py               # main src module
     calculator/src/calculator/forms.py
@@ -228,7 +223,7 @@ the applications that are installed. Note that in the following example, the val
 "--install" option is the path to the root directory of the application you want to install. In this
 example, it is a relative path to "/Users/philip/src/hue/calculator".
 
-        $ ./build/env/bin/python tools/app_reg/app_reg.py --install calculator --relative-paths
+        ./build/env/bin/python tools/app_reg/app_reg.py --install calculator --relative-paths
         === Installing app at calculator
         Updating registry with calculator (version 0.1)
         --- Making egg-info for calculator
@@ -257,7 +252,7 @@ Congrats, you've added a new app!
 You can now browse the new application.
 
     # If you haven't killed the old process, do so now.
-    $ build/env/bin/hue runserver
+    build/env/bin/hue runserver
 
 And then visit <a href="http://localhost:8000">http://localhost:8000/</a> to check it out!
 You should see the app (with a boring "SDK" icon) in the dock, and clicking it
@@ -301,7 +296,7 @@ to include a simple form:
 The template language here is <a href="http://www.makotemplates.org/docs/">Mako</a>,
 which is flexible and powerful.  If you use the "`.html`" extension, Hue
 will render your page using
-<a href="http://docs.djangoproject.com/en/1.2/topics/templates/#topics-templates">Django templates</a>
+<a href="https://docs.djangoproject.com/en/1.11/#the-template-layer">Django templates</a>
 instead.
 
 Note that we used the `url()` function to generate the URL to the calculator
@@ -322,13 +317,11 @@ Let's edit `calculator/src/calculator/views.py` to process that form:
         return render('index.mako', request, dict())
       a = float(request.REQUEST["a"])
       b = float(request.REQUEST["b"])
-      op = request.REQUEST["op"]
+      op = request.POST["op"]
       result = OPS[op](a, b)
       return render('index.mako', request,
         dict(a=a, b=b, op=OP_STRING[op], result=result))
 
-For more complicated forms, you may want to use Django Forms and
-avoid explicitly using `request.REQUEST`, but this is shorter.
 
 You can now go and try the calculator.  If you set everything up right, you
 should see something like:
@@ -336,13 +329,8 @@ should see something like:
 <img src="calculator_working.png">
 
 
-Integrate external Web applications in any language
-===================================================
-Use the [create_proxy_app command](http://gethue.com/integrate-external-web-applications-in-any-language)
-
-
-A Look at Three Existing Apps
-=============================
+A Look at some Existing Apps
+============================
 
 ![Arch](arch_examples.png)
 
@@ -362,53 +350,11 @@ No JavaScript was written to make this happen!  Instead, the template
 applied certain CSS classes to the relevant `div`'s, and JFrame
 did the rest.
 
-Proxy
------
-
-### Setup
-
-You need to have Hue running:
-
-    $ ./build/env/bin/hue runserver
-
-Then if you want to access localhost/50030/jobtracker.jsp you just do:
-
-    http://127.0.0.1:8000/proxy/localhost/50030/jobtracker.jsp
-
-and the page will be displayed within Hue.
-
-You can configure it in ``desktop/conf/pseudo-distributed.ini``
-
-    [proxy]
-    whitelist="(localhost|127\.0\.0\.1)50030|50070|50060|50075)",
-    #Comma-separated list of regular expressions, which match 'host:port' of requested proxy target.
 
-    blacklist=""
-    #Comma-separated list of regular expressions, which match any prefix of 'host:port/path' of requested proxy target.
-    # This does not support matching GET parameters.
-
-### Usage
-
-You can create a new app (or modify a current one for testing).
-
-Then in order to display the proxied page in your app, you could add in the template of a view of the new app a
-snippet of Javacript similar to this for loading the JobTracker page:
-
-    <script>
-        $.get('/proxy/localhost/50030/jobtracker.jsp', function(data) { $('#proxy-body').html(data); alert('Load was performed.'); });
-    </script>
-
-or alternatively get the page in the view (better solution) with the Hue
-[REST API](https://github.com/cloudera/hue/tree/master/desktop/core/src/desktop/lib/rest). Example of use of
-this API can be found in the [HDFS lib](https://github.com/cloudera/hue/blob/master/desktop/libs/hadoop/src/hadoop/fs/webhdfs.py).
-
-If you need to browse through the proxied page, using an iframe might be a better solution.
-
-
-Beeswax
--------
+SQL Hive Editor
+---------------
 
-Beeswax is on the opposite end of the complexity scale from Help.
+The Hive Editor (old named Beeswax) is on the opposite end of the complexity scale from Help.
 In addition to many views (in `views.py`), Beeswax uses
 Django Forms for server-side form validation (the forms are in `forms.py`),
 several features of the Mako templating engine (especially includes and
@@ -579,7 +525,7 @@ Walk-through of a Django View
 ![Django Flow](django_request.png)
 
 Django is an MVC framework, except that the controller is called a
-"[view](https://docs.djangoproject.com/en/1.2/topics/http/views/)" and
+"[view](https://docs.djangoproject.com/en/1.11/#the-view-layer)" and
 the "view" is called a "template".  For an application developer, the essential
 flow to understand is how the "urls.py" file provides a mapping between URLs (expressed as a
 regular expression, optionally with captured parameters) and view functions.
@@ -599,15 +545,15 @@ is:
       return render('view_function.mako', request, dict(greeting="hello"))
 
 The `render()` function chooses a template engine (either Django or Mako) based on the
-extension of the template file (".html" or ".mako").  Mako templates are more powerful,
+extension of the template file (".html" or ".mako"). Mako templates are more powerful,
 in that they allow you to run arbitrary code blocks quite easily, and are more strict (some
 would say finicky); Django templates are simpler, but are less expressive.
 
 Django Models
 -------------
 
-[Django Models](http://docs.djangoproject.com/en/1.2/topics/db/models/#topics-db-models)
-are Django's Object-Relational Mapping framework.  If your application
+[Django Models](https://docs.djangoproject.com/en/1.11/#the-model-layer)
+are Django's Object-Relational Mapping framework. If your application
 needs to store data (history, for example), models are a good way to do it.
 
 From an abstraction perspective, it's common to imagine external services
@@ -656,17 +602,12 @@ If you must use shared state, use Python's `threading.Lock`.
 
 Note that any module initialization may happen multiple times.
 Some WSGI containers (namely, Apache), will start multiple
-Unix processes, each with multiple threads.  So, while
+Unix processes, each with multiple threads. So, while
 you have to use locks to protect state within the process,
 there still may be multiple copies of this state.
 
 For persistent global state, it is common to place the state
-in the database.  If the state needs to be managed with application code,
-a common pattern to push state into a "helper process".  For example, in the Job Designer,
-a helper process keeps track of the processes that have been launched.  The Django views
-themselves are stateless, but they talk to this stateful helper process for
-updates.  A similar approach is taken with updating metrics for
-the Beeswax application.
+in the database or on the Browser local storage.
 
 Authentication Backends
 -----------------------
@@ -714,7 +655,7 @@ Profiling Hue Apps
 Hue has a profiling system built in, which can be used to analyze server-side
 performance of applications.  To enable profiling::
 
-    $ build/env/bin/hue runprofileserver
+    build/env/bin/hue runprofileserver
 
 Then, access the page that you want to profile.  This will create files like
 /tmp/useradmin.users.000072ms.2011-02-21T13:03:39.745851.prof.  The format for
@@ -761,7 +702,7 @@ application.
 
 In a nutshell, front-end development in Hue is using
 [Bootstrap](http://twitter.github.com/bootstrap/) and
-[jQuery](http://jquery.com/) to layout your app and script the custom
+[Knockout js](http://knockoutjs.com/) to layout your app and script the custom
 interactions.
 
 
@@ -850,48 +791,6 @@ These standard components have their own online documentation, which we will
 not repeat here. They let you write interactive behaviors with little or no
 JavaScript.
 
-## Key Differences from Hue 1.x
-
-Here are the key differences between the Hue 1.x front-end SDK and the later
-versions. In Hue 2.0 and beyond:
-
-* Since each page view only loads one application, you can declare HTML
-  elements by ID, you can declare and load your JavaScripts anywhere, and
-  are in full control of all the UI interactions.
-* You can use standard HTML links to other applications.
-* You do not need to register your application with the front-end, or declare
-  any dependencies using YAML.
-* The navigation bar is not pluggable in Hue 2.0.
-* The old "accordion" behavior can be replaced by
-  [Bootstrap collapse](http://twitter.github.com/bootstrap/javascript.html#collapse).
-* The old "art buttons" pattern can be replaced by [Bootstrap
-  buttons](http://twitter.github.com/bootstrap/base-css.html#buttons), and
-  [button
-  groups](http://twitter.github.com/bootstrap/components.html#buttonGroups).
-* The old "art inputs" pattern can be replaced by [Bootstrap
-  form inputs](http://twitter.github.com/bootstrap/base-css.html#forms).
-* The old "autocomplete" behavior can be replaced by [jQuery
-  autocomplete](http://jqueryui.com/demos/autocomplete/) or [Bootstrap
-  typeahead](http://twitter.github.com/bootstrap/javascript.html#typeahead).
-* The old "collapser" behavior can be replaced by
-  [Bootstrap collapse](http://twitter.github.com/bootstrap/javascript.html#collapse).
-* The old "context menu" behavior can be replaced by [Bootstrap button
-  dropdowns](http://twitter.github.com/bootstrap/components.html#buttonDropdowns).
-* The old "fittext" behavior is no longer supported.
-* The old "flash message" behavior is no longer supported.
-* The old "html table" behavior can be replaced by
-  [DataTables](http://datatables.net/).
-* The old "overtext" behavior can be replaced by [Bootstrap form
-  placeholder](http://twitter.github.com/bootstrap/base-css.html#forms).
-* The old "popup" behavior can be replaced by
-  [Bootstrap modals](http://twitter.github.com/bootstrap/javascript.html#modals).
-* The old "side-by-side select" pattern is no longer supported.
-* The old "splitview" layout is no longer supported.
-* The old "tabs" layout can be replaced by [Bootstrap
-  tabs](http://twitter.github.com/bootstrap/javascript.html#tabs).
-* The old "tool tips" behavior can be replaced by [Bootstrap
-  tooltips](http://twitter.github.com/bootstrap/javascript.html#tooltips).
-
 
 Including Other JavaScript Frameworks
 -------------------------------------
@@ -913,17 +812,17 @@ Internationalization
 ====================
 How to update all the messages and compile them::
 
-    $ make locales
+    make locales
 
 How to update and compile the messages of one app::
 
-    $ cd apps/beeswax
-    $ make compile-locale
+    cd apps/beeswax
+    make compile-locale
 
 How to create a new locale for an app::
 
-    $ cd $APP_ROOT/src/$APP_NAME/locale
-    $ $HUE_ROOT/build/env/bin/pybabel init -D django -i en_US.pot -d . -l fr
+    cd $APP_ROOT/src/$APP_NAME/locale
+    $HUE_ROOT/build/env/bin/pybabel init -D django -i en_US.pot -d . -l fr
 
 
 Debugging Tips and Tricks
@@ -946,17 +845,7 @@ Debugging Tips and Tricks
 
     `$ DESKTOP_DEPENDER_DEBUG=1 build/env/bin/hue runserver`
 
-* We highly recommend developing with the [Firebug](http://getfirebug.com)
-  debugging plugin for Firefox. With it enabled, you can use a utility called
-  [dbug](http://www.clientcide.com/docs/Core/dbug) which wraps Firebug
-  commands. This allows you to leave debug statements in your code and display
-  them on demand. In particular, typing in `dbug.cookie()` in Firebug will set
-  a cookie in your browser that will turn these statements on until you type
-  that command again to toggle them off. You'll see some of our own debugging
-  statements and you can add your own. In the future, entering this state may
-  also provide access to additional debugging features.
-* When the dbug state is enabled in the browser, right clicking on elements is
-  re-enabled which makes element inspection a little easier in Firebug.
+* We recommend developing with the Chrome console.
 
 <!--
 

+ 0 - 240
docs/user-guide/filebrowser.md

@@ -1,240 +0,0 @@
-
-<link rel="stylesheet" href="docbook.css" type="text/css" media="screen" title="no title" charset="utf-8"></link>
-
-File Browser
-============
-
-The File Browser application lets you browse and manipulate files and
-directories in the Hadoop Distributed File System (HDFS) while using
-Hue. With File Browser, you can:
-
--   Create files and directories, upload and download files, upload zip
-    archives, and rename, move, and delete files and directories. You
-    can also change a file's or directory's owner, group, and
-    permissions. See [Files and Directories](#filesAndDirectories).
--   Search for files, directories, owners, and groups. See [Searching
-    for Files and Directories](#searching).
--   View and edit files as text or binary. See [Viewing and Editing
-    Files](#viewAndEdit).
-
-File Browser Installation and Configuration
--------------------------------------------
-
-File Browser is one of the applications installed as part of Hue. For
-information about installing and configuring Hue, see the Hue Installation
-manual.
-
-Starting File Browser
----------------------
-
-Click the **File Browser** icon
-(![image](images/icon_filebrowser_24.png)) in the navigation bar at the
-top of the Hue browser page.
-
-
-<a id="fileAndDirectories"></a>
-Files and Directories
----------------------
-
-You can use File Browser to view the input and output files of your
-MapReduce jobs. Typically, you can save your output files in /tmp or in
-your home directory if your system administrator set one up for you. You
-must have the proper permissions to manipulate other user's files.
-
-### Creating Directories
-
-1.  In the File Browser window, select **New > Directory**.
-2.  In the **Create Directory** dialog box, enter a directory name and
-    then click **Submit**.
-
-### Changing Directories
-
--   Click the directory name or parent directory dots in the **File
-    Browser** window.
--   Click the ![image](images/edit.png) icon, type a directory name, and
-    press **Enter**.
-
-To change to your home directory, click **Home** in the path field at
-the top of the **File Browser** window.
-
-![image](images/note.jpg) **Note**:
-
-The **Home** button is disabled if you do not have a home directory. Ask
-a Hue administrator to create a home directory for you.
-
-### Creating Files
-
-1.  In the File Browser window, select **New > File**.
-2.  In the **Create File** dialog box, enter a file name and then click
-    **Submit**.
-
-
-<a id="uploadingFiles"></a>
-### Uploading Files
-
-You can upload text and binary files to the HDFS.
-
-1.  In the **File Browser** window, browse to the directory where you
-    want to upload the file.
-2.  Select **Upload \> Files**.
-3.  In the box that opens, click **Upload a File** to browse to and
-    select the file(s) you want to upload, and then click **Open**.
-
-### Copying a File
-
-1.  In the **File Browser** window, check the checkbox next to the file
-    you want to copy.
-2.  Click the ![image](images/copy.png) Copy button.
-
-### Downloading Files
-
-You can download text and binary files to the HDFS.
-
-1.  In the **File Browser** window, check the checkbox next to the file
-    you want to download.
-2.  Click the **Download** button.
-
-### Uploading Zip Archives
-
-You can upload zip archives to the HDFS. The archive is uploaded and
-extracted to a directory named archivename.
-
-1.  In the **File Browser** window, browse to the directory where you
-    want to upload the archive.
-2.  Select **Upload > Zip file**.
-3.  In the box that opens, click **Upload a zip file** to browse to and
-    select the archive you want to upload, and then click **Open**.
-
-### Trash Folder
-
-File Browser supports the HDFS trash folder (*home directory*/.Trash) to
-contain files and directories before they are permanently deleted. Files
-in the folder have the full path of the deleted files (in order to be
-able to restore them if needed) and checkpoints. The length of time a
-file or directory stays in the trash depends on HDFS properties.
-
-**Open**
-
-1.  In the **File Browser** window, click ![image](images/fbtrash.png).
-
-**Move Files and Directories To**
-
-1.  In the **File Browser** window, check the checkbox next to one or
-    more files and directories.
-2.  Select **Delete > Move to trash**.
-
-**Empty**
-
-1.  In the **File Browser** window, click ![image](images/fbtrash.png).
-2.  Click **Empty**.
-
-### Renaming, Moving, Deleting, and Restoring Files and Directories
-
-
-**Rename**
-
-1.  In the **File Browser** window, check the checkbox next to the file
-    or directory you want to rename.
-2.  Click the **Rename** button.
-3.  Enter the new name and then click **Submit**.
-
-**Move**
-
-1.  In the **File Browser** window, check the checkbox next to the file
-    or directory you want to move.
-2.  Click the **Move** button.
-3.  In the **Move** dialog box, browse to or type the new directory, and
-    then click **Submit**.
-
-**Delete**
-
-1.  In the **File Browser** window, check the checkbox next to the file
-    or directory you want to delete. If you select a directory, all of
-    the files and subdirectories contained within that directory are
-    also deleted.
-2.  Choose one of the following:
-    -   **Delete > Move to trash**
-    -   **Delete > Delete forever**
-
-3.  Click **Yes** to confirm. When you move a file to trash it is stored
-    in the .Trash folder in your home directory.
-
-**Restore**
-
-1.  In the **File Browser** window, open the .Trash folder.
-2.  Navigate to the folder containing the file you want to restore.
-3.  Check the checkbox next to the file.
-4.  Click **Restore**.
-
-### Changing a File's or Directory's Owner, Group, or Permissions
-
-![image](images/note.jpg) **Note**:
-
-Only the Hadoop superuser can change a file's or directory's owner,
-group, or permissions. The user who starts Hadoop is the Hadoop
-superuser. The Hadoop superuser account is not necessarily the same as a
-Hue superuser account. If you create a Hue user (in User Admin) with the
-same user name and password as the Hadoop superuser, then that Hue user
-can change a file's or directory's owner, group, or permissions.
-
-**Owner or Group**
-
-1.  In the **File Browser** window, check the checkbox next to the
-    select the file or directory whose owner or group you want to
-    change.
-2.  Choose **Change Owner/Group** from the Options menu.
-3.  In the **Change Owner/Group** dialog box:
-    -   Choose the new user from the **User** drop-down menu.
-    -   Choose the new group from the **Group** drop-down menu.
-    -   Check the **Recursive** checkbox to propagate the change.
-
-4.  Click **Submit** to make the changes.
-
-**Permissions**
-
-1.  In the **File Browser** window, check the checkbox next to the file
-    or directory whose permissions you want to change.
-2.  Click the **Change Permissions** button.
-3.  In the **Change Permissions** dialog box, select the permissions you
-    want to assign and then click **Submit**.
-    
-    
-<a id="searching"></a>
-Searching for Files and Directories
------------------------------------
-
-To search for files or directories by name using the query search box,
-enter the name of the file or directory in the query search box. File
-Browser lists the files or directories matching the search criteria.
-
-<a id="viewAndEdit"></a>
-Viewing and Editing Files
--------------------------
-
-You can view and edit files as text or binary.
-
-
-**View**
-
-1.  In the **File Browser** window, click the file you want to view.
-    File Browser displays the first 4,096 bytes of the file in the
-    **File Viewer** window.
-    -   If the file is larger than 4,096 bytes, use the Block navigation
-        buttons (First Block, Previous Block, Next Block, Last Block) to
-        scroll through the file block by block. The **Viewing Bytes**
-        fields show the range of bytes you are currently viewing.
-    -   To switch the view from text to binary, click **View as Binary**
-        to view a hex dump.
-    -   To switch the view from binary to text, click **View as Text**.
-
-**Edit**
-
-1.  If you are viewing a text file, click **Edit File**. File Browser
-    displays the contents of the file in the **File Editor** window.
-2.  Edit the file and then click **Save** or **Save As** to save the
-    file.
-
-**View Location in HDFS**
-
-Click **View File Location**. File Browser displays the file's location
-in the **File Browser** window.

+ 0 - 154
docs/user-guide/hbase.md

@@ -1,154 +0,0 @@
-
-<link rel="stylesheet" href="docbook.css" type="text/css" media="screen" title="no title" charset="utf-8"></link>
-
-HBase Browser
-=============
-
-We'll take a look at the new [HBase Browser App](http://gethue.com/the-web-ui-for-hbase-hbase-browser)
-added in Hue 2.5 and improved significantly since.
-
-Prerequisites before using the app:
-
-\1. Have HBase and Thrift Service 1 initiated (Thrift can be configured)
-
-\2. Configure your list of HBase Clusters in
-[hue.ini](https://github.com/cloudera/hue/blob/master/desktop/conf.dist/hue.ini#L467)
-to point to your Thrift IP/Port
-
-
-SmartView
----------
-
-The smartview is the view that you land on when you first enter a table.
-On the left hand side are the row keys and hovering over a row reveals a
-list of controls on the right. Click a row to select it, and once
-selected you can perform batch operations, sort columns, or do any
-amount of standard database operations. To explore a row, simple scroll
-to the right. By scrolling, the row should continue to lazily-load cells
-until the end.
-
-### Adding Data
-
-To initially populate the table, you can insert a new row or bulk upload
-CSV/TSV/etc. type data into your table.
-
-
-On the right hand side of a row is a '+' sign that lets you insert
-columns into your
-row
-
-### Mutating Data
-
-To edit a cell, simply click to edit inline.
-
-If you need more control or data about your cell, click “Full Editor” to
-edit.
-
-In the full editor, you can view cell history or upload binary data to
-the cell. Binary data of certain MIME Types are detected, meaning you
-can view and edit images, PDFs, JSON, XML, and other types directly in
-your browser!
-
-Hovering over a cell also reveals some more controls (such as the delete
-button or the timestamp). Click the title to select a few and do batch
-operations:
-
-If you need some sample data to get started and explore, check out this
-howto create [HBase table
-tutorial](http://gethue.com/hadoop-tutorial-how-to-create-example-tables-in-hbase).
-
-
-### Smart Searchbar
-
-The "Smart Searchbar" is a sophisticated tool that helps you zero-in on
-your data. The smart search supports a number of operations. The most
-basic ones include finding and scanning row keys. Here I am selecting
-two row keys with:
-
-
-    domain.100, domain.200
-
-
-Submitting this query gives me the two rows I was looking for. If I want
-to fetch rows after one of these, I have to do a scan. This is as easy
-as writing a '+' followed by the number of rows you want to fetch.
-
-
-    domain.100, domain.200 +5
-
-
-Fetches domain.100 and domain.200 followed by the next 5 rows. If you're
-ever confused about your results, you can look down below and the query
-bar and also click in to edit your query.
-
-The Smart Search also supports column filtering. On any row, I can
-specify the specific columns or families I want to retrieve. With:
-
-
-    domain.100[column_family:]   
-
-
-I can select a bare family, or mix columns from different families like
-so:
-
-
-    domain.100[family1:, family2:, family3:column_a]
-
-
-Doing this will restrict my results from one row key to the columns I
-specified. If you want to restrict column families only, the same effect
-can be achieved with the filters on the right. Just click to toggle a
-filter.
-
-
-Finally, let's try some more complex column filters. I can query for
-bare columns:
-
-
-    domain.100[column_a]
-
-This will multiply my query over all column families. I can also do
-prefixes and scans:
-
-
-    domain.100[family: prefix* +3]
-
-
-This will fetch me all columns that start with prefix\* limited to 3
-results. Finally, I can filter on range:
-
-
-    domain.100[family: column1 to column100]
-
-
-This will fetch me all columns in 'family:' that are lexicographically
-\>= column1 but <= column100. The first column ('column1') must be a
-valid column, but the second can just be any string for comparison.
-
-The Smart Search also supports prefix filtering on rows. To select a
-prefixed row, simply type the row key followed by a star \*. The prefix
-should be highlighted like any other searchbar keyword. A prefix scan is
-performed exactly like a regular scan, but with a prefixed row.
-
-
-    domain.10* +10
-
-
-Finally, as a new feature, you can also take full advantage of the
-[HBase filtering](denied:about:blank)language, by typing your filter
-string between curly braces. HBase Browser autocompletes your filters
-for you so you don't have to look them up every time. You can apply
-filters to rows or scans.
-
-
-    domain.1000 {ColumnPrefixFilter('100-') AND ColumnCountGetFilter(3)}
-
-
-This doc only covers a few basic features of the Smart Search. You can
-take advantage of the full querying language by referring to the help
-menu when using the app. These include column prefix, bare columns,
-column range, etc. Remember that if you ever need help with the
-searchbar, you can use the help menu that pops up while typing, which
-will suggest next steps to complete your query.
-
-

+ 0 - 210
docs/user-guide/impala.md

@@ -1,210 +0,0 @@
-
-<link rel="stylesheet" href="docbook.css" type="text/css" media="screen" title="no title" charset="utf-8"></link>
-
-Cloudera Impala Query UI
-========================
-
-The Cloudera Impala Query UI application enables you to perform queries
-on Apache Hadoop data stored in HDFS or HBase using Cloudera Impala. For
-information about Cloudera Impala, see [Installing and Using Cloudera
-Impala](/content/support/en/documentation/cloudera-impala/cloudera-impala-documentation-v1-latest.html).
-You can create, run, and manage queries, and download the results in a
-Microsoft Office Excel worksheet file or a comma-separated values file.
-
-Cloudera Impala Query UI Installation and Configuration
--------------------------------------------------------
-
-The Cloudera Impala Query UI application is one of the applications
-installed as part of Hue. For information about installing and
-configuring Hue, see the Hue Installation
-manual..
-
-The Cloudera Impala Query UI assumes an existing Cloudera Impala
-installation. The Hue installation instructions include the
-configuration necessary for Impala. You can view the current
-configuration from the **Settings** tab.
-
-Starting Cloudera Impala Query UI
----------------------------------
-
-Click the **Cloudera Impala Query UI** icon
-(![image](images/icon_impala_24.png)) in the navigation bar at the top
-of the Hue browser page.
-
-Managing Databases, Tables, and Partitions
-------------------------------------------
-
-You can create databases, tables, partitions, and load data by executing
-[Hive data manipulation
-statements](http://archive.cloudera.com/cdh4/cdh/4/hive/language_manual/data-manipulation-statements.html)
-in the Beeswax application.
-
-You can also use the [Metastore
-Manager]()
-application to manage the databases, tables, and partitions and load
-data.
-
-When you change the metastore using one of these applications, you must
-click the Refresh button under METASTORE CATALOG in the pane to the left
-of the Query Editor to make the metastore update visible to the Cloudera
-Impala server.
-
-Installing Example Queries and Tables
--------------------------------------
-
-![image](images/note.jpg) **Note**: You must be a superuser to perform
-this task.
-
-1.  Click ![image](images/quick_start.png). The Quick Start Wizard
-    opens.
-2.  Click **Step 2: Examples**.
-3.  Click **Cloudera Impala Query UI**.
-
-Query Editor
-------------
-
-The Query Editor view lets you create queries in the Cloudera Impala
-Query Language, which is based on the Hive Standard Query Language
-(HiveQL) and described in the Cloudera Impala Language Reference topic
-in [Installing and Using Cloudera
-Impala](http://www.cloudera.com/content/cloudera-content/cloudera-docs/Impala/latest/Installing-and-Using-Impala/Installing-and-Using-Impala.html).
-
-You can name and save your queries to use later.
-
-When you submit a query, you can either wait for the query to complete,
-or return later to find the queries in the **History** view.
-
-In the box to the left of the Query field, you can select a database,
-override the default Cloudera Impala settings, enable users to enter
-parameters at run-time. See [Advanced Query Settings](#advancedQuerySettings) for
-details on using these settings.
-
-### Creating Queries
-
-1.  In the Query Editor window, type a query or multiple queries
-    separated by a semicolon ";". To be presented with a drop-down of
-    autocomplete options, type CTRL+spacebar when entering a query.
-2.  To save your query and advanced settings to use again later, click
-    **Save As**, enter a name and description, and then click **OK**. To
-    save changes to an existing query, click **Save.**
-
-### Loading Queries into the Query Editor
-
-1.  Do one of the following:
-    -   Click the My Queries tab.
-        1.  Click the Recent Saved Queries or Recent Run Queries tab to
-            display the respective queries.
-
-    -   Click the Saved Queries tab.
-
-2.  Click a query name. The query is loaded into the Query Editor.
-
-### Running Queries
-
-![image](images/note.jpg) **Note**: To run a query, you must be logged
-in to Hue as a user that also has a Unix user account on the remote
-server.
-
-1.  To execute a portion of the query, highlight one or more query
-    statements.
-2.  Click **Execute**. The Query Results window appears with the results
-    of your query.
-    -   To view a log of the query execution, click **Log** at the top
-        of the results display. You can use the information in this tab
-        to debug your query.
-    -   To view the query that generated these results, click **Query**
-        at the top of the results display.
-    -   To view the columns of the query, click **Columns**.
-    -   To return to the query in the Query Editor, click **Unsaved
-        Query**.
-
-3.  If there are multiple statements in the query, click Next in the
-    Multi-statement query pane to execute the remaining statements.
-
-<a id="advancedQuerySettings"></a>
-### Advanced Query Settings
-
-The pane to the left of the Query Editor lets you specify the following
-options:
-
-<table>
-<tr><td>DATABASE</td><td>The database containing the table definitions.</td></tr>
-<tr><td>SETTINGS</td><td>Override the Cloudera Impala  default settings. To configure a new
-setting:
-
-<ol>
-<li> Click Add.
-<li> For Key, enter a Hive or Hadoop configuration variable name.
-<li> For Value, enter the value you want to use for the variable.
-
-For example, to override the directory where structured Hive query logs
-are created, you would enter hive.querylog.location for Key, and a
-path for Value.
-</ol>
-To view the default settings, click the Settings tab at the top of
-the page. 
-</td></tr>  
-<tr><td>PARAMETERIZATION</td><td>Indicate that a dialog box should display to enter parameter values when
-a query containing the string $parametername is executed. Enabled by
-default.</td></tr>
-<tr><td>METASTORE CATALOG</td><td>Refresh metadata. It is best to refresh metadata after making changes to
-databases such as adding or dropping a table.</td></tr>
-</table>
-
-
-### Viewing Query History
-
-You can view the history of queries that you have run previously.
-Results for these queries are available for one week or until Hue is
-restarted.
-
-1.  Click **History**. A list of your saved and unsaved queries displays
-    in the Query History window.
-2.  To display the queries for all users, click **Show everyone's
-    queries**. To display your queries only, click **Show my queries**.
-3.  To display the automatically generated actions performed on a user's
-    behalf, click **Show auto actions**. To display user queries again,
-    click **Show user queries**.
-
-### Viewing, Editing, Copying, and Deleting Saved Queries
-
-You can view a list of saved queries of all users by clicking **My
-Queries** and then selecting either Recent Saved Queries or Recent Run
-Queries tab to display the respective queries or clicking **Saved
-Queries**. You can copy any query, but you can edit, delete, and view
-the history of only your own queries.
-
-
-**Edit**
-
-1.  Click **Saved Queries**. The Queries window displays.
-2.  Check the checkbox next to the query and click **Edit**. The query
-    displays in the Query Editor window.
-3.  Change the query and then click **Save.** You can also click **Save
-    As**, enter a new name, and click **OK** to save a copy of the
-    query.
-
-**Copy**
-
-1.  Click **Saved Queries**. The Queries window displays.
-2.  Check the checkbox next to the query and click **Copy**. The query
-    displays in the Query Editor window.
-3.  Change the query as necessary and then click **Save.** You can also
-    click **Save As**, enter a new name, and click **OK** to save a copy
-    of the query.
-
-**Copy in Query History**
-
-1.  Click **History**. The Query History window displays.
-2.  To display the queries for all users, click **Show everyone's
-    queries**. The queries for all users display in the History window.
-3.  Click the query you want to copy. A copy of the query displays in
-    the Query Editor window.
-4.  Change the query, if necessary, and then click **Save As**, enter a
-    new name, and click **OK** to save the query.
-
-**Delete**
-
-1.  Click **Saved Queries**. The Queries window displays.
-2.  Check the checkbox next to the query and click **Delete**.
-3.  Click **Yes** to confirm the deletion.

+ 1812 - 290
docs/user-guide/index.md

@@ -3,7 +3,7 @@
 <link rel="stylesheet" href="bootplus.css" type="text/css" media="screen" title="no title" charset="utf-8"></link>
 
 
-# Hue User Guide
+<h1>Hue User Guide</h1>
 
 
 <div class="row-fluid">
@@ -14,290 +14,37 @@
    </div>
    <div class="span9">
 
-## Concept
+# Concept
 Hue consists in 4 apps in a single page interface that allow the users to perform its data
 analyzes without losing any context.
 
 Each app of Hue can be extended to support your own languages or apps.
 
-### Interface
-#### Top search
-#### Left assist
-#### Right assist
-#### Sample popup
-#### Charting
+## Interface
+### Top search
+### Left assist
+### Right assist
+### Sample popup
+### Charting
 The SDK will be clarified in the upcoming charting revamp.
 
-### Admin
-Quick Start Wizard
-------------------
-
-The Quick Start wizard allows you to perform the following Hue setup
-operations by clicking the tab of each step or sequentially by clicking
-Next in each screen:
-
-1.  **Check Configuration** validates your Hue configuration. It will
-    note any potential misconfiguration and provide hints as to how to
-    fix them. You can edit the configuration file described in the next
-    section or use Cloudera Manager, if installed, to manage your
-    changes.
-2.  **Examples** contains links to install examples into the Hive,
-    Impala, MapReduce, Spark, Oozie, Solr Dashboard and Pig Editor applications.
-3.  **Users** contains a link to the User Admin application to create or
-    import users and a checkbox to enable and disable collection of
-    usage information.
-4.  **Go!** - displays the Hue home screen, which contains links to the
-    different categories of applications supported by Hue: Query,
-    Hadoop, and Workflow.
-
-Configuration
--------------
-
-Displays a list of the installed Hue applications and their
-configuration. The location of the folder containing the Hue
-configuration files is shown at the top of the page. Hue configuration
-settings are in the hue.ini configuration file.
-
-Click the tabs under **Configuration Sections and Variables** to see the
-settings configured for each application. For information on configuring
-these settings, see Hue Configuration in the Hue installation manual.
-
-Server Logs
------------
-
-Displays the Hue Server log and allows you to download the log to your
-local system in a zip file.
-
-
-<link rel="stylesheet" href="docbook.css" type="text/css" media="screen" title="no title" charset="utf-8"></link>
-
-User Admin
-==========
-
-The User Admin application lets a superuser add, delete, and manage Hue
-users and groups, and configure group permissions. Superusers can add
-users and groups individually, or import them from an LDAP directory.
-Group permissions define the Hue applications visible to group members
-when they log into Hue and the application features available to them.
-
-Starting User Admin
--------------------
-
-Click the **User Admin** icon (![image](images/icon_useradmin_24.png))
-in the navigation bar at the top of the Hue browser page. The Hue Users
-page opens.
-
-Users
------
-
-The User Admin application provides two levels of user privileges:
-superusers and users.
-
--   Superusers — The first user who logs into Hue after its initial
-    installation becomes the first superuser. Superusers have
-    permissions to perform administrative functions:
-    -   Add and delete users
-    -   Add and delete groups
-    -   Assign permissions to groups
-    -   Change a user into a superuser
-    -   Import users and groups from an LDAP server
-
--   Users — can change their name, e-mail address, and password and log
-    in to Hue and run Hue applications, subject to the permissions
-    provided by the Hue groups to which they belong.
-
-### Adding a User
-
-1.  In the **User Admin** page, click **Add User**.
-2.  In the **Credentials** screen, add required information about the
-    user. Once you provide the required information you can click the
-    wizard step tabs to set other information.
-    
- <table>
-<tr><td>Username</td><td>  A user name that contains only letters, numbers, and underscores;
-    blank spaces are not allowed and the name cannot begin with a
-    number. The user name is used to log into Hue and in file
-    permissions and job submissions. This is a required field.
-</td></tr>
-<tr><td>Password and Password confirmation</td><td>    A password for the user. This is a required field.</td></tr>
-<tr><td>Create home directory</td><td>   Indicate whether to create a directory named /user/username in HDFS.
-    For non-superusers, the user and group of the directory are
-    username. For superusers, the user and group are username and
-    supergroup.</td></tr></table>
-
- 
-
-3.  Click **Add User** to save the information you specified and close
-    the **Add User** wizard or click **Next**.
-4.  In the **Names and Groups** screen, add optional information.
-
-<table>
-<tr><td>First name and Last name</td><td> The user's first and last name.
-</td></tr>
-<tr><td>E-mail address</td><td>The user's e-mail address. The e-mail address is used by the Job
-    Designer and Beeswax applications to send users an e-mail message
-    after certain actions have occurred. The Job Designer sends an
-    e-mail message after a job has completed. Beeswax sends a message
-    after a query has completed. If an e-mail address is not specified,
-    the application will not attempt to email the user.</td></tr>
-<tr><td>Groups</td><td> The groups to which the user belongs. By default, a user is assigned
-    to the **default** group, which allows access to all applications.
-    See [Permissions](#permissions).</td></tr></table>
-    
-
-5.  Click **Add User** to save the information you specified and close
-    the **Add User** wizard or click **Next**.
-6.  In the **Advanced** screen, add status information.
-
-<table>
-<tr><td>Active</td><td> Indicate that the user is enabled and allowed to log in. Default: checked.</td></tr>
-<tr><td>Superuser status</td><td> Assign superuser privileges to the user.</td></tr></table>
-
-7.  Click **Add User** to save the information you specified and close
-    the **Add User** wizard.
-
-### Deleting a User
-
-1.  Check the checkbox next to the user name and click **Delete**.
-2.  Click **Yes** to confirm.
-
-### Editing a User
-
-1.  Click the user you want to edit in the **Hue Users** list.
-2.  Make the changes to the user and then click **Update user**.
-
-### Importing Users from an LDAP Directory
-
-Hue must be configured to use an external LDAP directory (OpenLDAP or
-Active Directory). See Hue Installation in [CDH4
-Installation](http://www.cloudera.com/content/cloudera-content/cloudera-docs/CDH4/latest/CDH4-Installation-Guide/CDH4-Installation-Guide.html).
-
-![image](images/note.jpg) **Note**:
-
-Importing users from an LDAP directory does not import any password
-information. You must add passwords manually in order for a user to log
-in.
-
-To add a user from an external LDAP directory:
-
-1.  Click **Add/sync LDAP user**.
-2.  Specify the user properties:
-
-<table>
-<tr><td>Username</td><td>The user name.</td></tr>
-<tr><td>Distinguished name</td><td>Indicate that Hue should use a full distinguished name for the user.
-    This imports the user's first and last name, username, and email,
-    but does not store the user password.</td></tr>
-    <tr><td>Create home directory</td><td> Indicate that Hue should create a home directory for the user in
-    HDFS.</td></tr></table>
-
-
-3.  Click **Add/sync user**.
-
-    If the user already exists in the User Admin, the user information
-    in User Admin is synced with what is currently in the LDAP
-    directory.
-
-### Syncing Users and Groups with an LDAP Directory
-
-You can sync the Hue user database with the current state of the LDAP
-directory using the **Sync LDAP users/groups** function. This updates
-the user and group information for the already imported users and
-groups. It does not import any new users or groups.
-
-1.  Click **Sync LDAP users/groups**.
-2.  The **Create Home Directories** checkbox creates home directories in
-    HDFS for existing imported members that don't have home directories.
-3.  In the **Sync LDAP users and groups** dialog, click **Sync** to
-    perform the sync.
-
-Groups
-------
-
-Superusers can add and delete groups, configure group permissions, and
-assign users to group memberships.
-
-### Adding a Group
-
-You can add groups, and delete the groups you've added. You can also
-import groups from an LDAP directory.
-
-1.  In the **User Admin** window, click **Groups** and then click **Add
-    Group**.
-2.  Specify the group properties:
-
-<table>
-<tr><td>Name</td><td> The name of the group. Group names can only be letters, numbers, and
-    underscores; blank spaces are not allowed.</td></tr>
-<tr><td>Members</td><td>The users in the group. Check user names or check Select all.</td></tr>
-    <tr><td>Permissions</td><td>The applications the users in the group can access. Check
-    application names or check Select all.</td></tr></table>
-
-3.  Click **Add group**.
-
-### Adding Users to a Group
-
-1.  In the **User Admin** window, click **Groups**.
-2.  Click the group.
-3.  To add users to the group, check the names in the list provided or
-    check **Select All**.
-4.  Click **Update group**.
-
-### Deleting a Group
-
-1.  Click **Groups**.
-2.  Check the checkbox next to the group and click **Delete**.
-3.  Click **Yes** to confirm.
-
-### Importing Groups from an LDAP Directory
-
-1.  From the **Groups** tab, click **Add/sync LDAP group**.
-2.  Specify the group properties:
-
-<table>
-<tr><td>Name</td><td> The name of the group.</td></tr>
-<tr><td>Distinguished name</td><td> Indicate that Hue should use a full distinguished name for the
-    group.</td></tr>
-    <tr><td>Import new members</td><td>  Indicate that Hue should import the members of the group.</td></tr>
-        <tr><td>Import new members from all subgroups</td><td>
-    Indicate that Hue should import the members of the subgroups.</td></tr>
-            <tr><td>Create home directories</td><td> Indicate that Hue should create home directories in HDFS for the
-    imported members.</td></tr></table>
-
-3.  Click **Add/sync group**.
-
-<a id="permissions"></a>
-Permissions
------------
-
-Permissions for Hue applications are granted to groups, with users
-gaining permissions based on their group membership. Group permissions
-define the Hue applications visible to group members when they log into
-Hue and the application features available to them.
-
-1.  Click **Permissions**.
-2.  Click the application for which you want to assign permissions.
-3.  Check the checkboxes next to the groups you want to have permission
-    for the application. Check **Select all** to select all groups.
-4.  Click **Update permission**. The new groups will appear in the
-    Groups column in the **Hue Permissions** list.
-
-
-### Importer
+## Importer
+### Tables
+### Indexes
+### DBMS
 
-### Documents
-#### Sharing
-#### Import / Export
+## Documents
+### Sharing
+### Import / Export
 
-## Editors
+# Editors
 The goal of Hue’s Editor is to make data querying easy and productive.
 
 It focuses on SQL but also supports job submissions. It comes with an intelligent autocomplete, search & tagging of data and query assistance.
 
-#### Languages
-#### Hive
-Beeswax
-=======
+## Languages
+### Hive
+### Impala
 
 The Beeswax application enables you to perform queries on Apache Hive, a
 data warehousing system designed to work with Hadoop. For information
@@ -307,8 +54,6 @@ create Hive databases, tables and partitions, load data, create, run,
 and manage queries, and download the results in a Microsoft Office Excel
 worksheet file or a comma-separated values file.
 
-Beeswax and Hive Installation and Configuration
------------------------------------------------
 
 Beeswax is installed and configured as part of Hue. For information
 about installing and configuring Hue, see the Hue Installation
@@ -436,29 +181,555 @@ a query containing the string $parametername is executed. Enabled by
 default.</td></tr>
 </table>
 
-#### Autocomplete
-#### Syntax checker
-#### Query Assist
-#### Risk Alerts
-#### Presentation
-#### SDK
+## Autocomplete
+## Syntax checker
+## Query Assist
+## Risk Alerts
+## Presentation
 
+### Jobs
 
-## Dashboards
-Dashboards are an interactive way to explore your data quickly and easily. No programming is required and the analysis is done by drag & drops and clicks.
+The Job Designer application enables you to create and submit jobs to
+the Hadoop cluster. You can include variables with your jobs to enable
+you and other users to enter values for the variables when they run your
+job. The Job Designer supports the actions supported by
+[Oozie](http://archive.cloudera.com/cdh4/cdh/4/oozie/):
+[MapReduce](/content/cloudera-content/cloudera-docs/HadoopTutorial/CDH4/index.html),
+Streaming, Java, Pig, Hive, Sqoop, Shell, Ssh, DistCp, Fs, and Email.
 
-#### SDK
+Job Designer Installation and Configuration
+-------------------------------------------
 
+Job Designer is one of the applications installed as part of Hue. For
+information about installing and configuring Hue, see the Hue Installation
+manual..
 
-## Browsers
-Hue’s Browsers powers your Data Catalog. They let you easily search, glance and perform actions on data or jobs in Cloud or on premise clusters.
+In order to run DistCp, Streaming, Pig, Sqoop, and Hive jobs, Oozie must
+be configured to use the Oozie ShareLib. See the Oozie Installation manual.
 
-### File Browser
+Starting Job Designer
+---------------------
 
-<link rel="stylesheet" href="docbook.css" type="text/css" media="screen" title="no title" charset="utf-8"></link>
+Click the **Job Designer** icon (![image](images/icon_jobsub_24.png)) in
+the navigation bar at the top of the Hue web page. The **Job Designs**
+page opens in the browser.
+
+Installing the Example Job Designs
+----------------------------------
+
+![image](images/note.jpg) **Note**: You must be a superuser to perform
+this task.
+
+1.  Click ![image](images/quick_start.png). The Quick Start Wizard
+    opens.
+2.  Click **Step 2: Examples**.
+3.  Click **Job Designer**.
+
+Job Designs
+-----------
+
+A job design specifies several meta-level properties of a job, including
+the job design name, description, the executable scripts or classes, and
+any parameters for those scripts or classes.
+
+### Filtering Job Designs
+
+You can filter the job designs that appear in the list by owner, name,
+type, and description.
+
+**To filter the Job Designs list:**
+
+1.  In the **Job Designs** window, click **Designs**.
+2.  Enter text in the Filter text box at the top of the **Job Designs**
+    window. When you type in the Filter field, the designs are
+    dynamically filtered to display only those rows containing text that
+    matches the specified substring.
+
+### Creating a Job Design
+
+1.  In the **Job Designs** window, click New Action \> Action, where
+    Action is MapReduce, Streaming, Java, Pig, Hive, Sqoop, Shell, Ssh,
+    DistCp, Fs, or Email.
+2.  In the **Job Design (Action type)** window, specify the common and
+    job type specific information.
+3.  Click **Save** to save the job settings.
+
+### Deleting and Restoring Job Designs
+
+You can move job designs to the trash and later restore or permanently
+delete them.
+
+#### Deleting Job Designs
+
+1.  In a Manager screen, check the checkbox next to one or more job
+    designs.
+2.  Choose one of the following:
+    -   Delete \> Move to trash
+    -   Delete \> Delete forever
+
+#### Restoring Job Designs
+
+1.  In a Manager screen, click ![image](images/trash.png) **Trash**.
+2.  Check the checkbox next to one or more job designs.
+3.  Click Restore.
+
+### Job Design Settings
+
+#### Job Design Common Settings
+
+Most job design types support all the settings listed in the following
+table. For job type specific settings, see:
+[MapReduce](#mapreduce),
+[Streaming](#streaming),
+[Java](#java),
+[Pig](#pig),
+[Hive](#hive),
+[Sqoop](#sqoop),
+[Shell](#shell),
+[Ssh](#ssh),
+[DistCp](#distcp),
+[Fs](#fs), and
+[Email](#email).
+
+All job design settings except Name and Description support the use of
+variables of the form $variable\_name. When you run the job, a dialog
+box will appear to enable you to specify the values of the variables.
+
+<table>
+<tr><td>Name</td><td>Identifies the job and its collection of properties and parameters.</td></tr>
+<tr><td>Description</td><td>A description of the job. The description is displayed in the dialog box
+that appears if you specify variables for the job.</td></tr>
+<tr><td>Advanced</td><td>Advanced settings:<ul><li>Is shared- Indicate whether to share the action with all users.<li>Oozie parameters - parameters to pass to Oozie</td></tr>
+<tr><td>Prepare</td><td>Specifies paths to create or delete before starting the workflow job.</td></tr>
+<tr><td>Params</td>Parameters to pass to a script or command. The parameters are expressed
+using the [JSP 2.0 Specification (JSP.2.3) Expression
+Language](http://jcp.org/aboutJava/communityprocess/final/jsr152/),
+allowing variables, functions, and complex expressions as parameters.<td></td></tr>
+<tr><td>Job Properties</td><td>Job properties. To set a property value, click <b>Add Property</b>.<ol><li>Property name -  a configuration property name. This field provides autocompletion, so you can type the first few characters of a property name and then select the one you want from the drop-down
+    list.<li>Valuethe property value.</td></tr>
+<tr><td>Files</td><td>Files to pass to the job. Equivalent to the Hadoop -files option.</td></tr>
+<tr><td>Archives</td><td>Files to pass to the job. Archives to pass to the job. Equivalent to the Hadoop -archives option.</td></tr></table>
+
+<a id="mapreduce"></a>
+#### MapReduce Job Design
+
+A MapReduce job design consists of MapReduce functions written in Java.
+You can create a MapReduce job design from existing mapper and reducer
+classes without having to write a main Java class. You must specify the
+mapper and reducer classes as well as other MapReduce properties in the
+Job Properties setting.
+
+<table>
+<tr><td>Jar path</td><td>The fully-qualified path to a JAR file containing the classes that
+implement the Mapper and Reducer functions.</td></tr>
+</table>
+
+<a id="streaming"></a>
+#### Streaming Job Design
+
+Hadoop streaming jobs enable you to create MapReduce functions in any
+non-Java language that reads standard Unix input and writes standard
+Unix output. For more information about Hadoop streaming jobs, see
+[Hadoop
+Streaming](http://archive.cloudera.com/cdh/3/hadoop-0.20.2+320/streaming.html).
+
+<table>
+<tr><td>Mapper</td><td>The path to the mapper script or class. If the mapper file is not on the
+machines on the cluster, use the Files option to pass it as a part
+of job submission. Equivalent to the Hadoop -mapper option.</td></tr>
+<tr><td>Reducer</td><td>The path to the reducer script or class. If the reducer file is not on
+the machines on the cluster, use the Files option to pass it as a
+part of job submission. Equivalent to the Hadoop -reducer option.</td></tr>
+</table>
+
+<a id="java"></a>
+#### Java Job Design
+
+A Java job design consists of a main class written in Java.
+
+<table>
+<tr><td>Jar path</td><td>The fully-qualified path to a JAR file containing the main class.</td></tr>
+<tr><td>Main class</td><td>The main class to invoke the program.</td></tr>
+<tr><td>Args</td><td>The arguments to pass to the main class.</td></tr>
+<tr><td>Java opts</td><td>The options to pass to the JVM.</td></tr>
+</table>
+
+<a id="pig"></a>
+#### Pig Job Design
+
+
+A Pig job design consists of a Pig script.
+
+<table>
+<tr><td>Script name</td><td>Script name or path to the Pig script.</td></tr>
+</table>
+
+<a id="hive"></a>
+#### Hive Job Design
+
+A Hive job design consists of a Hive script.
+
+<table>
+<tr><td>Script name</td><td>Script name or path to the Hive script.</td></tr>
+</table>
+
+
+<a id="sqoop"></a>
+#### Sqoop Job Design
+
+A Sqoop job design consists of a Sqoop command.
+
+<table>
+<tr><td>Command</td><td>The Sqoop command.</td></tr>
+</table>
+
+<a id="shell"></a>
+#### Shell Job Design
+
+A Shell job design consists of a shell command.
+
+<table>
+<tr><td>Command</td><td>The shell command.</td></tr>
+<tr><td></td>Capture output<td>Indicate whether to capture the output of the command.</td></tr>
+</table>
+
+<a id="ssh"></a>
+#### Ssh Job Design
+
+A Ssh job design consists of an ssh command.
+
+<table>
+<tr><td>User</td><td>The name of the user to run the command as.</td></tr>
+<tr><td>Host</td><td>The name of the host to run the command on.</td></tr>
+<tr><td>Command</td><td>The ssh command.</td></tr>
+<tr><td></td>Capture output<td>Indicate whether to capture the output of the command.</td></tr>
+</table>
+
+<a id="distcp"></a>
+#### DistCp Job Design
+
+A DistCp job design consists of a DistCp command.
+
+<a id="fs"></a>
+#### Fs Job Design
+
+A Fs job design consists of a command that operates on HDFS.
+
+<table>
+<tr><td>Delete path</td><td>The path to delete. If it is a directory, it deletes recursively all its
+content and then deletes the directory.</td></tr>
+<tr><td></td>Create directory<td>The path of a directory to create.</td></tr>
+<tr><td>Move file</td><td>The source and destination paths to the file to be moved.</td></tr>
+<tr><td>Change permissions</td><td>The path whose permissions are to be changed, the permissions, and an
+indicator of whether to change permission recursively.</td></tr></table>
+
+<a id="email"></a>
+#### Email Job Design
+
+A Email job design consists of an email message.
+
+<table>
+<tr><td>To addresses</td><td>The recipient of the email message.</td></tr>
+<tr><td>CC addresses (optional)</td><td>The cc recipients of the email message.</td></tr>
+<tr><td>Subject</td><td>The subject of the email message.</td></tr>
+<tr><td>Body</td><td>The body of the email message.</td></tr>
+</table>
+
+
+### Submitting a Job Design
+
+![image](images/note.jpg) **Note**:
+
+A job's input files must be uploaded to the cluster before you can
+submit the job.
+
+**To submit a job design:**
+
+1.  In the **Job Designs** window, click **Designs** in the upper left
+    corner. Your jobs and other users' jobs are displayed in the **Job
+    Designs** window.
+2.  Check the checkbox next to the job you want to submit.
+3.  Click the **Submit** button.
+    1.  If the job contains variables, enter the information requested
+        in the dialog box that appears. For example, the sample grep
+        MapReduce design displays a dialog where you specify the output
+        directory.
+    2.  Click **Submit** to submit the job.
+
+After the job is complete, the Job Designer displays the results of the
+job. For information about displaying job results, see [Displaying the
+Results of Submitting a Job](#submitJob).
+
+### Copying, Editing, and Deleting a Job Design
+
+If you want to edit and use a job but you don't own it, you can make a
+copy of it and then edit and use the copied job.
+
+
+**Copy**
+
+1.  In the **Job Designs** window, click **Designs**. The jobs are
+    displayed in the **Job Designs** window.
+2.  Check the checkbox next to the job you want to copy.
+3.  Click the **Copy** button.
+4.  In the **Job Design Editor** window, change the settings and then
+    click **Save** to save the job settings.
+
+**Edit**
+
+1.  In the **Job Designs** window, click **Designs**. The jobs are
+    displayed in the **Job Designs** window.
+2.  Check the checkbox next to the job you want to edit.
+3.  Click the **Edit** button.
+4.  In the **Job Design** window, change the settings and then click
+    **Save** to save the job settings.
+
+Delete
+
+1.  In the **Job Designs** window, click **Designs**. The jobs are
+    displayed in the **Job Designs** window.
+2.  Check the checkbox next to the job you want to delete.
+3.  Click the **Delete** button.
+4.  Click **OK** to confirm the deletion.
+
+<a id="submitJob"></a>
+Displaying Results of Submitting a Job
+--------------------------------------
+
+**To display the Job Submission History:**
+
+In the **Job Designs** window, click the **History** tab. The jobs are
+displayed in the **Job Submissions History** listed by Oozie job ID.
+
+**To display Job Details:**
+
+In the **Job Submission History** window, click an Oozie Job ID. The
+results of the job display:
+
+-   Actions - a list of actions in the job.
+-   Click ![image](images/gear.png) to display the action configuration.
+    In the action configuration for a MapReduce action, click the value
+    of the mapred.output.dir property to display the job output.
+-   In the root-node row, click the Id in the External Id column to view
+    the job in the Job Browser.
+-   Details - the job details. Click ![image](images/gear.png) to
+    display the Oozie application configuration.
+-   Definition - the Oozie application definition.
+-   Log - the output log.
+
+## SDK
+
+
+# Dashboards
+Dashboards are an interactive way to explore your data quickly and easily. No programming is required and the analysis is done by drag & drops and clicks.
+
+### SDK
+
+
+# Browsers
+Hue’s Browsers powers your Data Catalog. They let you easily search, glance and perform actions on data or jobs in Cloud or on premise clusters.
+
+## Tables
+
+The Table Browser enables you to manage the databases,
+tables, and partitions of the metastore shared by
+the ([Beeswax](../beeswax.html) and [Cloudera Impala Query
+UI](../impala.html)) applications. You can use Metastore
+Manager to perform the following operations:
+
+-   Databases
+    -   [Select a database](#selectDatabase)
+    -   [Create a database](#createDatabase)
+    -   [Drop databases](#dropDatabase)
+
+-   Tables
+    -   [Create tables](#createTables)
+    -   [Browse tables](#browseTables)
+    -   [Import data into a table](#importDataIntoTables)
+    -   [Drop tables](#dropTables)
+    -   [View the location of a table](#viewTableLocation)
+
+Metastore Manager Installation and Configuration
+------------------------------------------------
+
+Metastore Manager is one of the applications installed as part of Hue.
+For information about installing and configuring Hue, see the Hue Installation
+manual.
+
+Starting Metastore Manager
+--------------------------
+
+Click the **Metastore Manager** icon
+(![image](images/icon_table_browser_24.png)) in the navigation bar at
+the top of the Hue browser page.
+
+### Installing Sample Tables
+
+![image](images/note.jpg) **Note**: You must be a superuser to perform
+this task.
+
+1.  Click ![image](images/quick_start.png). The Quick Start Wizard
+    opens.
+2.  Click **Step 2: Examples**.
+3.  Click **Beeswax (Hive UI)** or **Cloudera Impala Query UI**.
+
+### Importing Data
+
+If you want to import your own data instead of installing the sample
+tables, follow the procedure in [Creating Tables](#createTables).
+
+<a id="selectDatabase"></a>
+Selecting a Database
+--------------------
+
+1.  In the pane on the left, select the database from the DATABASE
+    drop-down list.
+
+<a id="createDatabase"></a>
+Creating a Database
+-------------------
 
-File Browser
-============
+1.  Click ![image](images/databases.png).
+2.  Click **Create a new database**.
+    1.  Specify a database name and optional description. Database names
+        are not case-sensitive. Click **Next**.
+    2.  Do one of the following:
+        -   Keep the default location in the Hive warehouse folder.
+        -   Specify an external location within HDFS:
+            1.  Uncheck the **Location** checkbox.
+            2.  In the External location field, type a path to a folder
+                on HDFS or click ![image](images/browse.png) to browse
+                to a folder and click **Select this folder**.
+
+    3.  Click the **Create Database** button.
+    
+<a id="selectDatabase"></a>
+Dropping Databases
+------------------
+
+1.  Click ![image](images/databases.png).
+2.  In the list of databases, check the checkbox next to one or more
+    databases.
+3.  Click the ![image](images/trash.png) Drop button.
+4.  Confirm whether you want to delete the databases.
+
+<a id="createTables"></a>
+Creating Tables
+---------------
+
+Although you can create tables by executing the appropriate Hive HQL DDL
+query commands, it is easier to create a table using the Metastore
+Manager table creation wizard.
+
+There are two ways to create a table: from a file or manually. If you
+create a table from a file, the format of the data in the file will
+determine some of the properties of the table, such as the record and
+file formats. The data from the file you specify is imported
+automatically upon table creation. When you create a file manually, you
+specify all the properties of the table, and then execute the resulting
+query to actually create the table. You then import data into the table
+as an additional step.
+
+**From a File**
+
+1.  In the ACTIONS pane in the Metastore Manager window, click **Create
+    a new table from a file**. The table creation wizard starts.
+2.  Follow the instructions in the wizard to create the table. The basic
+    steps are:
+    -   Choose your input file. The input file you specify must exist.
+        Note that you can choose to have Beeswax create the table
+        definition only based on the import file you select, without
+        actually importing data from that file.
+    -   Specify the column delimiter.
+    -   Define your columns, providing a name and selecting the type.
+
+3.  Click **Create Table** to create the table. The new table's metadata
+    displays on the right side of the **Table Metadata** window. At this
+    point, you can view the metadata or a sample of the data in the
+    table. From the ACTIONS pane you can import new data into the table,
+    browse the table, drop it, or go to the File Browser to see the
+    location of the data.
+
+**Manually**
+
+1.  In the ACTIONS pane in the Metastore Manager window, click **Create
+    a new table manually**. The table creation wizard starts.
+2.  Follow the instructions in the wizard to create the table. The basic
+    steps are:
+    -   Name the table.
+    -   Choose the record format.
+    -   Configure record serialization by specifying delimiters for
+        columns, collections, and map keys.
+    -   Choose the file format.
+    -   Specify the location for your table's data.
+    -   Specify the columns, providing a name and selecting the type for
+        each column.
+    -   Specify partition columns, providing a name and selecting the
+        type for each column.
+
+3.  Click **Create table**. The Table Metadata window displays.
+
+<a id="browseTables"></a>
+Browsing Tables
+---------------
+
+**To browse table data:**
+
+In the Table List window, check the checkbox next to a table name and
+click **Browse Data**. The table's data displays in the Query Results
+window.
+
+**To browse table metadata:**
+
+Do one of the following:
+
+-   In the Table List window, click a table name.
+-   Check the checkbox next to a table name and click **View**.
+
+-   The table's metadata displays in the **Columns** tab. You can view
+    the table data by selecting the **Sample** tab.
+-   If the table is partitioned, you can view the partition columns by
+    clicking the **Partition Columns** tab and display the partitions by
+    clicking **Show Partitions(n)**, where n is the number of partitions
+    in the ACTIONS pane on the left.
+
+<a id="importDataIntoTables"></a>
+Importing Data into a Table
+---------------------------
+
+When importing data, you can choose to append or overwrite the table's
+data with data from a file.
+
+1.  In the Table List window, click the table name. The Table Metadata
+    window displays.
+2.  In the ACTIONS pane, click **Import Data**.
+3.  For **Path**, enter the path to the file that contains the data you
+    want to import.
+4.  Check **Overwrite existing data** to replace the data in the
+    selected table with the imported data. Leave unchecked to append to
+    the table.
+5.  Click **Submit**.
+
+<a id="dropTables"></a>
+Dropping Tables
+---------------
+
+1.  In the Table List window, click the table name. The Table Metadata
+    window displays.
+2.  In the ACTIONS pane, click **Drop Table**.
+3.  Click **Yes** to confirm the deletion.
+
+<a id="viewTableLocation"></a>
+Viewing a Table's Location
+--------------------------
+
+1.  In the Table List window, click the table name. The Table Metadata
+    window displays.
+2.  Click **View File Location**. The file location of the selected
+    table displays in its directory in the File Browser window.
+
+
+## File Browser
 
 The File Browser application lets you browse and manipulate files and
 directories in the Hadoop Distributed File System (HDFS), S3 or ADLS.
@@ -473,6 +744,11 @@ With File Browser, you can:
 -   View and edit files as text or binary. See [Viewing and Editing
     Files](#viewAndEdit).
 
+### File systems
+#### HDFS
+#### S3
+#### ADLS
+
 File Browser Installation and Configuration
 -------------------------------------------
 
@@ -690,9 +966,1255 @@ You can view and edit files as text or binary.
 Click **View File Location**. File Browser displays the file's location
 in the **File Browser** window.
 
+Job Browser
+===========
+
+The Job Browser application lets you to examine the Hadoop MapReduce
+jobs running on your Hadoop cluster. Job Browser presents the job and
+tasks in layers. The top layer is a list of jobs, and you can link to a
+list of that job's tasks. You can then view a task's attempts and the
+properties of each attempt, such as state, start and end time, and
+output size. To troubleshoot failed jobs, you can also view the logs of
+each attempt.
+
+Job Browser Installation and Configuration
+------------------------------------------
+
+Job Browser is one of the applications installed as part of Hue. For
+information about installing and configuring Hue, see the Hue Installation
+manual.
 
-#### Scheduler
+Job Browser can display both MRv1 and MRv2 jobs, but must be configured
+to display one type at a time. 
+
+Starting Job Browser
+--------------------
+
+Click the **Job Browser** icon (![image](images/icon_jobbrowser_24.png))
+in the navigation bar at the top of the Hue web page.
+
+If there are no jobs that have been run, the **Welcome to the Job
+Browser** page opens, with links to the Job Designer and Beeswax.
+
+If there are jobs running, then the Job Browser list appears.
+
+Filtering the Job Browser List
+------------------------------
+
+-   To filter the jobs by their state (such as **Running** or
+    **Completed**), choose a state from the **Job status** drop-down
+    menu.
+-   To filter by a user who ran the jobs, enter the user's name in the
+    **User Name** query box.
+-   To filter by job name, enter the name in the **Text** query box.
+-   To clear the filters, choose **All States** from the **Job status**
+    drop-down menu and delete any text in the **User Name** and **Text**
+    query boxes.
+-   To display retired jobs, check the **Show retired jobs** checkbox.
+    Retired jobs show somewhat limited information – for example,
+    information on maps and reduces and job duration is not available.
+    Jobs are designated as Retired by the JobTracker based on the value
+    of mapred.jobtracker.retirejob.interval. The retired jobs no longer
+    display after the JobTracker is restarted.
+
+Viewing Job Information and Logs
+--------------------------------
+
+![image](images/note.jpg) **Note**: At any level you can view the log
+for an object by clicking the ![image](images/log.png) icon in the Logs
+column.
+
+**To view job information for an individual job:**
+
+1.  In the **Job Browser** window, click **View** at the right of the
+    job you want to view. This shows the **Job** page for the job, with
+    the recent tasks associated with the job are displayed in the
+    **Tasks** tab.
+2.  Click the **Metadata** tab to view the metadata for this job.
+3.  Click the **Counters** tab to view the counter metrics for the job.
+
+**To view details about the tasks associated with the job:**
+
+1.  In the **Job** window, click the **View All Tasks** link at the
+    right just above the **Recent Tasks** list. This lists all the tasks
+    associated with the job.
+2.  Click **Attempts** to the right of a task to view the attempts for
+    that task.
+
+**To view information about an individual task:**
+
+1.  In the **Job** window, click the **View** link to the right of the
+    task. The attempts associated with the task are displayed.
+2.  Click the **Metadata** tab to view metadata for this task. The
+    metadata associated with the task is displayed.
+3.  To view the Hadoop counters for a task, click the **Counters** tab.
+    The counters associated with the task are displayed.
+4.  To return to the **Job** window for this job, click the job number
+    in the status panel at the left of the window.
+
+**To view details about a task attempt:**
+
+1.  In the **Job Task** window, click the **View** link to the right of
+    the task attempt. The metadata associated with the attempt is
+    displayed under the **Metadata** tab.
+2.  To view the Hadoop counters for the task attempt, click the
+    **Counters** tab. The counters associated with the attempt are
+    displayed.
+3.  To view the logs associated with the task attempt, click the
+    **Logs** tab. The logs associated with the task attempt are
+    displayed.
+4.  To return to the list of tasks for the current job, click the task
+    number in the status panel at the left of the window.
+
+
+### Scheduler
 The application lets you build workflows and then schedule them to run regularly automatically. A monitoring interface shows the progress, logs and allow actions like pausing or stopping jobs.
 
+
+The Oozie Editor/Dashboard application allows you to define Oozie
+workflow, coordinator, and bundle applications, run workflow,
+coordinator, and bundle jobs, and view the status of jobs. For
+information about Oozie, see [Oozie
+Documentation](http://archive.cloudera.com/cdh4/cdh/4/oozie/).
+
+A workflow application is a collection of actions arranged in a directed
+acyclic graph (DAG). It includes two types of nodes:
+
+-   Control flow - start, end, fork, join, decision, and kill
+-   Action - [MapReduce](../jobdesigner.html#mapreduce),
+[Streaming](../jobdesigner.html#streaming),
+[Java](../jobdesigner.html#java),
+[Pig](../jobdesigner.html#pig),
+[Hive](../jobdesigner.html#hive),
+[Sqoop](../jobdesigner.html#sqoop),
+[Shell](../jobdesigner.html#shell),
+[Ssh](../jobdesigner.html#ssh),
+[DistCp](../jobdesigner.html#distcp),
+[Fs](../jobdesigner.html#fs), and
+[Email](../jobdesigner.html#email).
+    In order to run DistCp, Streaming, Pig, Sqoop, and Hive jobs, Oozie
+    must be configured to use the Oozie ShareLib. See the Oozie Installation
+   manual.
+
+A coordinator application allows you to define and execute recurrent and
+interdependent workflow jobs. The coordinator application defines the
+conditions under which the execution of workflows can occur.
+
+A bundle application allows you to batch a set of coordinator
+applications.
+
+Oozie Editor/Dashboard Installation and Configuration
+-----------------------------------------------------
+
+Oozie Editor/Dashboard is one of the applications installed as part of
+Hue. For information about installing and configuring Hue, see the Hue Installation
+manual.
+
+Starting Oozie Editor/Dashboard
+-------------------------------
+
+Click the **Oozie Editor/Dashboard** icon
+(![image](images/icon_oozie_24.png)) in the navigation bar at the top of
+the Hue browser page. **Oozie Editor/Dashboard** opens with the
+following screens:
+
+-   [Dashboard](#dashboard) - shows the running and completed workflow,
+    coordinator, and bundle jobs and information about Oozie
+    instrumentation and configuration. The screen is selected and opened
+    to the Workflows page.
+-   [Workflow Manager](#workflowManager) - shows available workflows and
+    allows you to create and import workflows.
+-   [Coordinator Manager](#coordinatorManager) - shows available coordinators and
+    allows you to create coordinators.
+-   [Bundle Manager](#bundleManager) - shows available bundles and
+    allows you to create bundles.
+
+Installing Oozie Editor/Dashboard Examples
+------------------------------------------
+
+![image](images/note.jpg) **Note**: You must be a superuser to perform
+this task.
+
+1.  Click ![image](images/quick_start.png). The Quick Start Wizard
+    opens.
+2.  Click **Step 2: Examples**.
+3.  Click **Oozie Editor/Dashboard**.
+
+Filtering Lists in Oozie Editor/Dashboard
+-----------------------------------------
+
+Many screens contain lists. When you type in the Filter field on
+screens, the lists are dynamically filtered to display only those rows
+containing text that matches the specified substring.
+
+Permissions in Oozie Editor/Dashboard
+-------------------------------------
+
+In the Dashboard workflows, coordinators, and bundles can only be
+viewed, submitted, and modified by their owner or a superuser.
+
+Editor permissions for performing actions on workflows, coordinators,
+and bundles are summarized in the following table:
+
+<table>
+<th><td>Action</td><td>Superuser or Owner</td><td>All</td></th>
+<tr><td>View</td><td>Y.</td><td>Only if "Is shared" is set</td></tr>
+<tr><td>Submit</td><td>Y.</td><td>Only if "Is shared" is set</td></tr>
+<tr><td>Modify</td><td>Y.</td><td>N</td></tr>
+</table>
+
+
+Deleting and Restoring Workflows, Coordinators, and Bundles
+-----------------------------------------------------------
+
+You can move workflows, coordinators, and bundles to the trash and later
+restore or permanently delete them.
+
+### Deleting Workflows, Coordinators, and Bundles
+
+1.  In a Manager screen, check the checkbox next to one or more
+    workflows, coordinators or bundles.
+2.  Choose one of the following:
+    -   Delete \> Move to trash
+    -   Delete \> Delete forever
+
+### Restoring Workflows, Coordinators, and Bundles
+
+1.  In a Manager screen, click ![image](images/trash.png) **Trash**.
+2.  Check the checkbox next to one or more workflows, coordinators or
+    bundles.
+3.  Click Restore.
+
+
+<a id="dashboard"></a>
+Dashboard
+---------
+
+The Dashboard shows a summary of the running and completed workflow,
+coordinator, and bundle jobs.
+
+You can view jobs for a period up to the last 30 days.
+
+You can filter the list by date (1, 7, 15, or 30 days) or status
+(Succeeded, Running, or Killed). The date and status buttons are
+toggles.
+
+
+### Workflows
+
+Click the **Workflows** tab to view the running and completed workflow
+jobs for the filters you have specified.
+
+Click a workflow row in the Running or Completed table to view detailed
+information about that workflow job.
+
+In the left pane contains a link to the workflow and the variable values
+specified.![image](images/workflow.jpg)
+
+For the selected job, the following information is available in the
+right area.
+
+-   **Graph** tab shows the workflow DAG.
+-   **Actions** tab shows you details about the actions that make up the
+    workflow.
+    -   Click the **Id** link to see additional details about the
+        action.
+    -   Click the **External Id** link to view the job in the Job
+        Browser.
+
+-   **Details** tab shows job statistics including start and end times.
+-   **Configuration** tab shows selected job configuration settings.
+-   **Logs** tab shows log output generated by the workflow job.
+-   **Definition** tab shows the Oozie workflow definition, as it
+    appears in the workflow.xml file (also linked under the application
+    path properties in the **Details** tab and the **Configuration**
+    tab).
+
+For each action in the workflow you can:
+
+-   Click the ![image](images/eye.png) icon to view the action screen,
+    which contains:
+    -   **Details** tab shows job statistics including start and end
+        times.
+    -   **Configuration** tab shows the action configuration settings.
+    -   **Child Jobs** tab lists jobs generated by the action.
+
+-   Click the ![image](images/log.png) icon to view the log in the Job
+    Browser.
+
+
+### Coordinators
+
+Click the **Coordinators** tab to view the running and completed
+coordinator jobs for the filters you have specified.
+
+For the selected job, the following information is available.
+
+-   The **Calendar** tab shows the timestamp of the job. Click the
+    timestamp to open the workflow DAG.
+-   The **Actions** tab shows you details about the actions that make up
+    the coordinator.
+    -   Click the **Id** link to see additional details about the
+        action.
+    -   Click the **External Id** link to view the job in the Job
+        Browser.
+
+-   The **Configuration** tab shows selected job configuration settings.
+-   The **Logs** tab shows log output generated by the coordinator.
+-   The **Definition** tab shows the Oozie coordinator definition, as it
+    appears in the coordinator.xml file (also linked under the
+    oozie.coord.application.path property in the **Configuration** tab).
+
+### Bundles
+
+Click the **Bundles** tab to view the running and completed bundle jobs
+for the filters you have specified.
+
+### Oozie
+
+The Oozie tab provides subtabs that give you access to Oozie
+instrumentation and configuration settings.
+
+#### Instrumentation
+
+For information on the instrumentation metrics supported by Oozie, see
+[Oozie
+Monitoring](http://oozie.apache.org/docs/3.3.0/AG_Monitoring.html).
+
+#### Configuration
+
+For information on the configuration properties supported by Oozie, see
+[Oozie
+Configuration](http://oozie.apache.org/docs/3.3.0/AG_Install.html#Oozie_Configuration).
+
+<a id="workflowManager"></a>
+Workflow Manager
+----------------
+
+In Workflow Manager you create Oozie workflows and submit them for
+execution.
+
+Click the **Workflows** tab to open the Workflow Manager.
+
+Each row shows a workflow: its name, description, timestamp of its last
+modification. It also shows:
+
+-   **Steps** - the number of steps in the workflow execution path. This
+    is the number of execution steps between the start and end of the
+    workflow. This will not necessarily be the same as the number of
+    actions in the workflow, if there are control flow nodes in the
+    control path.
+-   **Status** - who can run the workflow. **shared** means users other
+    than the owner can access the workflow. **personal** means only the
+    owner can modify or submit the workflow. The default is personal.
+-   **Owner** - the user that created the workflow.
+
+In Workflow Editor you edit workflows that include MapReduce, Streaming,
+Java, Pig, Hive, Sqoop, Shell, Ssh, DistCp, Fs, Email, Sub-workflow, and
+Generic actions. You can configure these actions in the Workflow Editor,
+or you can import job designs from Job Designer to be used as actions in
+your workflow. For information about defining workflows, see the
+[Workflow
+Specification](http://archive.cloudera.com/cdh4/cdh/4/oozie/WorkflowFunctionalSpec.html).
+
+### Opening a Workflow
+
+To open a workflow, in Workflow Manager, click the workflow. Proceed
+with [Editing a Workflow](#editingWorkflow).
+
+### Creating a Workflow
+
+1.  Click the **Create** button at the top right.
+2.  In the Name field, type a name.
+3.  Check the Is shared checkbox to allow all users to access the
+    workflow.
+4.  Click **advanced** to specify the deployment directory or a job.xml
+    file.
+5.  Click **Save**. The Workflow Editor opens. Proceed with [Editing a
+    Workflow](#editingWorkflow).
+
+### Importing a Workflow
+
+1.  Click the **Import** button at the top right.
+2.  In the Name field, type a name.
+3.  In the **Local workflow.xml file** field, click **Choose File** and
+    select a workflow file.
+4.  Click **advanced** to specify whether the workflow is shared, the
+    deployment directory, or a job.xml file.
+5.  Click **Save**. The Workflow Editor opens. Proceed with [Editing a
+    Workflow](#editingWorkflow).
+
+### Submitting a Workflow
+
+To submit a workflow for execution, do one of the following:
+
+-   In the Workflow Manager, click the radio button next to the
+    workflow, and click the **Submit** button.
+-   In the Workflow Editor, click the **Submit** button.
+
+The workflow job is submitted and the Dashboard displays the workflow
+job.
+
+To view the output of the job, click ![image](images/log.png) **View the
+logs**.
+
+#### Suspending a Running Job
+
+In the pane on the left, click the **Suspend** button.
+
+1.  Verify that you want to suspend the job.
+
+#### Resuming a Suspended Job
+
+In the pane on the left, click the **Resume** button.
+
+1.  Verify that you want to resume the job.
+
+#### Rerunning a Workflow
+
+In the pane on the left, click the **Rerun** button.
+
+1.  Check the checkboxes next to the actions to rerun.
+2.  Specify required variables.
+3.  Click **Submit**.
+
+### Scheduling a Workflow
+
+To schedule a workflow for recurring execution, do one of the following:
+
+-   In the Workflow Manager, check the checkbox next to the workflow and
+    click the **Schedule** button.
+-   In the Workflow Editor, click the **Schedule** button.
+
+A coordinator is created and opened in the Coordinator Editor. Proceed
+with [Editing a Coordinator](#editingCoordinator).
+
+<a id="editingWorkflow"></a>
+### Editing a Workflow
+
+In the Workflow Editor you can easily perform operations on Oozie action
+and control nodes.
+
+#### Action Nodes
+
+The Workflow Editor supports dragging and dropping action nodes. As you
+move the action over other actions and forks, highlights indicate active
+areas. If there are actions in the workflow, the active areas are the
+actions themselves and the areas above and below the actions. If you
+drop an action on an existing action, a fork and join is added to the
+workflow.
+
+-   Add actions to the workflow by clicking an action
+    ![image](images/action.png) button and drop the action on the
+    workflow. The Edit Node screen displays.
+    1.  Set the action properties and click **Done**. Each action in a
+        workflow must have a unique name.
+
+-   Copy an action by clicking the ![image](images/copy.png) **Copy**
+    button.
+
+1.  The action is opened in the Edit Node screen.
+2.  Edit the action properties and click **Done**. The action is added
+    to the end of the workflow.
+
+-   Delete an action by clicking the ![image](images/trash.png) button.
+-   Edit an action by clicking the ![image](images/edit.png) button.
+-   Change the position of an action by left-clicking and dragging an
+    action to a new location.
+
+#### Control Nodes
+
+-   Create a fork and join by dropping an action on top of another
+    action.
+-   Remove a fork and join by dragging a forked action and dropping it
+    above the fork.
+-   Convert a fork to a decision by clicking the
+    ![image](images/convert.png) button.
+-   To edit a decision:
+    1.  Click the ![image](images/edit.png) button.
+    2.  Fill in the predicates that determine which action to perform
+        and select the default action from the drop-down list.
+    3.  Click **Done**.
+
+### Uploading Workflow Files
+
+In the Workflow Editor, click the **Upload** button.
+
+The workspace of the workflow is opened in the File Browser application.
+Follow the procedure in [Uploading
+Files](../filebrowser.html#uploadingFiles) to upload the files. You must
+put JAR files in a lib directory in the workspace.
+
+### Editing Workflow Properties
+
+1.  In the Workflow Editor, click the link under the Name or Description
+    fields in the left pane.
+2.  To share the workflow with all users, check the **Is shared**
+    checkbox.
+3.  To set advanced execution options, click **advanced** and edit the
+    deployment directory, add parameters and job properties, or specify
+    a job.xml file.
+4.  Click **Save**.
+
+### Displaying the History of a Workflow
+
+1.  Click the **Dashboard** tab.
+2.  Click the **Workflows** tab.
+3.  Click a workflow.
+4.  Click the **Actions** tab.
+
+<a id="coordinatorManager"></a>
+Coordinator Manager
+-------------------
+
+In Coordinator Manager you create Oozie coordinator applications and
+submit them for execution.
+
+Click the **Coordinators** tab to open the Coordinator Manager.
+
+Each row shows a coordinator: its name, description, timestamp of its
+last modification. It also shows:
+
+-   **Workflow** - the workflow that will be run by the coordinator.
+-   **Frequency** - how often the workflow referenced by the coordinator
+    will be run.
+-   **Status** - who can run the coordinator. **shared** means users
+    other than the owner can access the workflow. **personal** means
+    only the owner can modify or submit the workflow. The default is
+    personal.
+-   **Owner** - the user that created the coordinator.
+
+In Coordinator Editor, you edit coordinators and the datasets required
+by the coordinators. For information about defining coordinators and
+datasets, see the [Coordinator
+Specification](http://archive.cloudera.com/cdh4/cdh/4/oozie/CoordinatorFunctionalSpec.html).
+
+### Opening a Coordinator
+
+To open a coordinator, in Coordinator Manager, click the coordinator.
+Proceed with [Editing a Coordinator](#editingCoordinator).
+
+### Creating a Coordinator
+
+To create a coordinator, in Coordinator Manager:
+
+1.  Click the **Create** button at the top right. The Coordinator wizard
+    opens. Proceed with [Editing a Coordinator](#editingCoordinator).
+
+### Submitting a Coordinator
+
+To submit a coordinator for execution, check the checkbox next to the
+coordinator and click the **Submit** button.
+
+<a id="editingCoordinator"></a>
+### Editing a Coordinator
+
+In the Coordinator Editor you specify coordinator properties and the
+datasets on which the workflow scheduled by the coordinator will operate
+by stepping through screens in a wizard. You can also advance to
+particular steps and revisit steps by clicking the Step "tabs" above the
+screens. The following instructions walk you through the wizard.
+
+1.  Type a name, select the workflow, check the **Is shared checkbox**
+    to share the job, and click **Next**. If the Coordinator Editor was
+    opened after scheduling a workflow, the workflow will be set.
+2.  Select how many times the coordinator will run for each specified
+    unit, the start and end times of the coordinator, the timezone of
+    the start and end times, and click **Next**. Times must be expressed
+    as UTC times. For example, to run at 10 pm PST, specify a start time
+    of 6 am UTC of the following day (+8 hours) and set the Timezone
+    field to America/Los\_Angeles.
+3.  Click **Add** to select an input dataset and click **Next**. If no
+    datasets exist, follow the procedure in [Creating a
+    Dataset](#creatingDataset).
+4.  Click **Add** to select an output dataset. Click **Save
+    coordinator** or click **Next** to specify advanced settings.
+5.  To share the coordinator with all users, check the **Is****shared**
+    checkbox.
+6.  Fill in parameters to pass to Oozie, properties that determine how
+    long a coordinator will wait before timing out, how many
+    coordinators can run and wait concurrently, and the coordinator
+    execution policy.
+7.  Click **Save coordinator**.
+
+<a id="creatingDataset"></a>
+### Creating a Dataset
+
+1.  In the Coordinator Editor, do one of the following:
+    -   Click **here** in the Inputs or Outputs pane at the top of the
+        editor.
+    -   In the pane at the left, click the **Create new** link. Proceed
+        with [Editing a Dataset](#editingDataset).
+
+### Displaying Datasets
+
+1.  In the Coordinator Editor, click **Show existing** in pane at the
+    left.
+2.  To edit a dataset, click the dataset name in the Existing datasets
+    table. Proceed with [Editing a Dataset](#editingDataset).
+
+<a id="editingDataset"></a>
+### Editing a Dataset
+
+1.  Type a name for the dataset.
+2.  In the Start and Frequency fields, specify when and how often the
+    dataset will be available.
+3.  In the URI field, specify a URI template for the location of the
+    dataset. To construct URIs and URI paths containing dates and
+    timestamps, you can specify the variables
+    ${YEAR},${MONTH},${DAY},${HOUR},${MINUTE}. For example:
+    hdfs://foo:9000/usr/app/stats/${YEAR}/${MONTH}/data.
+4.  In the Instance field, click a button to choose a default, single,
+    or range of data instances. For example, if frequency==DAY, a window
+    of the last rolling 5 days (not including today) would be expressed
+    as start: -5 and end: -1. Check the advanced checkbox to display a
+    field where you can specify a coordinator [EL
+    function](http://archive.cloudera.com/cdh4/cdh/4/oozie/CoordinatorFunctionalSpec.html).
+5.  Specify the timezone of the start date.
+6.  In the Done flag field, specify the flag that identifies when input
+    datasets are no longer ready.
+
+### Displaying the History of a Coordinator
+
+1.  Click the **Dashboard** tab.
+2.  Click the **Coordinators** tab.
+3.  Click a coordinator.
+4.  Click the **Actions** tab.
+
+<a id="bundleManager"></a>
+Bundle Manager
+--------------
+
+In Bundle Manager you create Oozie bundle applications and submit them
+for execution.
+
+Click the **Bundle** tab to open the Bundle Manager.
+
+Each row shows a bundle: its name, description, timestamp of its last
+modification. It also shows:
+
+-   **Coordinators** - the coordinators that will be run by the bundle.
+-   **Kick off** - the UTC time when the coordinators referenced by the
+    bundle will be started.
+-   **Status** - who can run the bundle. **shared** means users other
+    than the owner can access the workflow. **personal** means only the
+    owner can modify or submit the workflow. The default is personal.
+-   **Owner** - the user that created the bundle.
+
+For information about defining bundles, see the [Bundle
+Specification](http://archive.cloudera.com/cdh4/cdh/4/oozie/BundleFunctionalSpec.html).
+
+### Opening a Bundle
+
+To open a bundle, in Bundle Manager, click the bundle. Proceed with
+[Editing a Bundle](#editingBundle).
+
+### Creating a Bundle
+
+1.  Click the **Create** button at the top right.
+2.  In the Name field, type a name.
+3.  In the Kick off time field, choose a kick off time.
+4.  Check the Is shared checkbox to allow all users to access the
+    workflow.
+5.  Click **Save**. The Bundle Editor opens. Proceed with [Editing a
+    Bundle](#editingBundle).
+
+### Submitting a Bundle
+
+To submit a bundle for execution, check the checkbox next to the bundle
+and click the **Submit** button.
+
+<a id="editingBundle"></a>
+### Editing a Bundle
+
+In the Bundle Editor, you specify properties by stepping through screens
+in a wizard. You can also advance to particular steps and revisit steps
+by clicking the Step "tabs" above the screens. The following
+instructions walk you through the wizard.
+
+1.  Click **Add** to select a coordinator that the bundle will kick off.
+2.  Choose the kick off time. The time must be expressed as a UTC time.
+    For example, to run at 10 pm PST, specify a start time of 6 am UTC
+    of the following day (+8 hours).
+3.  To share the bundle with all users, check the **Is shared**
+    checkbox.
+4.  Click **Next** to specify advanced settings or click **Save
+    bundle**.
+5.  Fill in parameters to pass to Oozie.
+6.  Click **Save bundle**.
+
+Displaying the History of a Bundle
+----------------------------------
+
+1.  Click the **Dashboard** tab.
+2.  Click the **Bundles** tab.
+3.  Click a bundle.
+4.  Click the **Actions** tab.
+
+
+
+# Administration
+
+Quick Start Wizard
+------------------
+
+The Quick Start wizard allows you to perform the following Hue setup
+operations by clicking the tab of each step or sequentially by clicking
+Next in each screen:
+
+1.  **Check Configuration** validates your Hue configuration. It will
+    note any potential misconfiguration and provide hints as to how to
+    fix them. You can edit the configuration file described in the next
+    section or use Cloudera Manager, if installed, to manage your
+    changes.
+2.  **Examples** contains links to install examples into the Hive,
+    Impala, MapReduce, Spark, Oozie, Solr Dashboard and Pig Editor applications.
+3.  **Users** contains a link to the User Admin application to create or
+    import users and a checkbox to enable and disable collection of
+    usage information.
+4.  **Go!** - displays the Hue home screen, which contains links to the
+    different categories of applications supported by Hue: Query,
+    Hadoop, and Workflow.
+
+Configuration
+-------------
+
+Displays a list of the installed Hue applications and their
+configuration. The location of the folder containing the Hue
+configuration files is shown at the top of the page. Hue configuration
+settings are in the hue.ini configuration file.
+
+Click the tabs under **Configuration Sections and Variables** to see the
+settings configured for each application. For information on configuring
+these settings, see Hue Configuration in the Hue installation manual.
+
+Server Logs
+-----------
+
+Displays the Hue Server log and allows you to download the log to your
+local system in a zip file.
+
+
+## User management
+
+The User Admin application lets a superuser add, delete, and manage Hue
+users and groups, and configure group permissions. Superusers can add
+users and groups individually, or import them from an LDAP directory.
+Group permissions define the Hue applications visible to group members
+when they log into Hue and the application features available to them.
+
+Starting User Admin
+-------------------
+
+Click the **User Admin** icon (![image](images/icon_useradmin_24.png))
+in the navigation bar at the top of the Hue browser page. The Hue Users
+page opens.
+
+Users
+-----
+
+The User Admin application provides two levels of user privileges:
+superusers and users.
+
+-   Superusers — The first user who logs into Hue after its initial
+    installation becomes the first superuser. Superusers have
+    permissions to perform administrative functions:
+    -   Add and delete users
+    -   Add and delete groups
+    -   Assign permissions to groups
+    -   Change a user into a superuser
+    -   Import users and groups from an LDAP server
+
+-   Users — can change their name, e-mail address, and password and log
+    in to Hue and run Hue applications, subject to the permissions
+    provided by the Hue groups to which they belong.
+
+### Adding a User
+
+1.  In the **User Admin** page, click **Add User**.
+2.  In the **Credentials** screen, add required information about the
+    user. Once you provide the required information you can click the
+    wizard step tabs to set other information.
+    
+ <table>
+<tr><td>Username</td><td>  A user name that contains only letters, numbers, and underscores;
+    blank spaces are not allowed and the name cannot begin with a
+    number. The user name is used to log into Hue and in file
+    permissions and job submissions. This is a required field.
+</td></tr>
+<tr><td>Password and Password confirmation</td><td>    A password for the user. This is a required field.</td></tr>
+<tr><td>Create home directory</td><td>   Indicate whether to create a directory named /user/username in HDFS.
+    For non-superusers, the user and group of the directory are
+    username. For superusers, the user and group are username and
+    supergroup.</td></tr></table>
+
+ 
+
+3.  Click **Add User** to save the information you specified and close
+    the **Add User** wizard or click **Next**.
+4.  In the **Names and Groups** screen, add optional information.
+
+<table>
+<tr><td>First name and Last name</td><td> The user's first and last name.
+</td></tr>
+<tr><td>E-mail address</td><td>The user's e-mail address. The e-mail address is used by the Job
+    Designer and Beeswax applications to send users an e-mail message
+    after certain actions have occurred. The Job Designer sends an
+    e-mail message after a job has completed. Beeswax sends a message
+    after a query has completed. If an e-mail address is not specified,
+    the application will not attempt to email the user.</td></tr>
+<tr><td>Groups</td><td> The groups to which the user belongs. By default, a user is assigned
+    to the **default** group, which allows access to all applications.
+    See [Permissions](#permissions).</td></tr></table>
+    
+
+5.  Click **Add User** to save the information you specified and close
+    the **Add User** wizard or click **Next**.
+6.  In the **Advanced** screen, add status information.
+
+<table>
+<tr><td>Active</td><td> Indicate that the user is enabled and allowed to log in. Default: checked.</td></tr>
+<tr><td>Superuser status</td><td> Assign superuser privileges to the user.</td></tr></table>
+
+7.  Click **Add User** to save the information you specified and close
+    the **Add User** wizard.
+
+### Deleting a User
+
+1.  Check the checkbox next to the user name and click **Delete**.
+2.  Click **Yes** to confirm.
+
+### Editing a User
+
+1.  Click the user you want to edit in the **Hue Users** list.
+2.  Make the changes to the user and then click **Update user**.
+
+### Importing Users from an LDAP Directory
+
+Hue must be configured to use an external LDAP directory (OpenLDAP or
+Active Directory). See Hue Installation in [CDH4
+Installation](http://www.cloudera.com/content/cloudera-content/cloudera-docs/CDH4/latest/CDH4-Installation-Guide/CDH4-Installation-Guide.html).
+
+![image](images/note.jpg) **Note**:
+
+Importing users from an LDAP directory does not import any password
+information. You must add passwords manually in order for a user to log
+in.
+
+To add a user from an external LDAP directory:
+
+1.  Click **Add/sync LDAP user**.
+2.  Specify the user properties:
+
+<table>
+<tr><td>Username</td><td>The user name.</td></tr>
+<tr><td>Distinguished name</td><td>Indicate that Hue should use a full distinguished name for the user.
+    This imports the user's first and last name, username, and email,
+    but does not store the user password.</td></tr>
+    <tr><td>Create home directory</td><td> Indicate that Hue should create a home directory for the user in
+    HDFS.</td></tr></table>
+
+
+3.  Click **Add/sync user**.
+
+    If the user already exists in the User Admin, the user information
+    in User Admin is synced with what is currently in the LDAP
+    directory.
+
+### Syncing Users and Groups with an LDAP Directory
+
+You can sync the Hue user database with the current state of the LDAP
+directory using the **Sync LDAP users/groups** function. This updates
+the user and group information for the already imported users and
+groups. It does not import any new users or groups.
+
+1.  Click **Sync LDAP users/groups**.
+2.  The **Create Home Directories** checkbox creates home directories in
+    HDFS for existing imported members that don't have home directories.
+3.  In the **Sync LDAP users and groups** dialog, click **Sync** to
+    perform the sync.
+
+Groups
+------
+
+Superusers can add and delete groups, configure group permissions, and
+assign users to group memberships.
+
+### Adding a Group
+
+You can add groups, and delete the groups you've added. You can also
+import groups from an LDAP directory.
+
+1.  In the **User Admin** window, click **Groups** and then click **Add
+    Group**.
+2.  Specify the group properties:
+
+<table>
+<tr><td>Name</td><td> The name of the group. Group names can only be letters, numbers, and
+    underscores; blank spaces are not allowed.</td></tr>
+<tr><td>Members</td><td>The users in the group. Check user names or check Select all.</td></tr>
+    <tr><td>Permissions</td><td>The applications the users in the group can access. Check
+    application names or check Select all.</td></tr></table>
+
+3.  Click **Add group**.
+
+### Adding Users to a Group
+
+1.  In the **User Admin** window, click **Groups**.
+2.  Click the group.
+3.  To add users to the group, check the names in the list provided or
+    check **Select All**.
+4.  Click **Update group**.
+
+### Deleting a Group
+
+1.  Click **Groups**.
+2.  Check the checkbox next to the group and click **Delete**.
+3.  Click **Yes** to confirm.
+
+### Importing Groups from an LDAP Directory
+
+1.  From the **Groups** tab, click **Add/sync LDAP group**.
+2.  Specify the group properties:
+
+<table>
+<tr><td>Name</td><td> The name of the group.</td></tr>
+<tr><td>Distinguished name</td><td> Indicate that Hue should use a full distinguished name for the
+    group.</td></tr>
+    <tr><td>Import new members</td><td>  Indicate that Hue should import the members of the group.</td></tr>
+        <tr><td>Import new members from all subgroups</td><td>
+    Indicate that Hue should import the members of the subgroups.</td></tr>
+            <tr><td>Create home directories</td><td> Indicate that Hue should create home directories in HDFS for the
+    imported members.</td></tr></table>
+
+3.  Click **Add/sync group**.
+
+<a id="permissions"></a>
+Permissions
+-----------
+
+Permissions for Hue applications are granted to groups, with users
+gaining permissions based on their group membership. Group permissions
+define the Hue applications visible to group members when they log into
+Hue and the application features available to them.
+
+1.  Click **Permissions**.
+2.  Click the application for which you want to assign permissions.
+3.  Check the checkboxes next to the groups you want to have permission
+    for the application. Check **Select all** to select all groups.
+4.  Click **Update permission**. The new groups will appear in the
+    Groups column in the **Hue Permissions** list.
+
+
+# Contrib
+Those modules are not active enough to be officially maintained in the core Hue but can still
+fit your needs. Any [contribution](https://github.com/cloudera/hue/wiki) is still welcomed!
+
+## SDK
+Check the SDK guide or contact the community about how to build your own custom app.
+
+## HBase Browser
+
+We'll take a look at the new [HBase Browser App](http://gethue.com/the-web-ui-for-hbase-hbase-browser).
+
+Prerequisites before using the app:
+
+\1. Have HBase and Thrift Service 1 initiated (Thrift can be configured)
+
+\2. Configure your list of HBase Clusters in
+[hue.ini](https://github.com/cloudera/hue/blob/master/desktop/conf.dist/hue.ini#L467)
+to point to your Thrift IP/Port
+
+
+SmartView
+---------
+
+The smartview is the view that you land on when you first enter a table.
+On the left hand side are the row keys and hovering over a row reveals a
+list of controls on the right. Click a row to select it, and once
+selected you can perform batch operations, sort columns, or do any
+amount of standard database operations. To explore a row, simple scroll
+to the right. By scrolling, the row should continue to lazily-load cells
+until the end.
+
+### Adding Data
+
+To initially populate the table, you can insert a new row or bulk upload
+CSV/TSV/etc. type data into your table.
+
+
+On the right hand side of a row is a '+' sign that lets you insert
+columns into your
+row
+
+### Mutating Data
+
+To edit a cell, simply click to edit inline.
+
+If you need more control or data about your cell, click “Full Editor” to
+edit.
+
+In the full editor, you can view cell history or upload binary data to
+the cell. Binary data of certain MIME Types are detected, meaning you
+can view and edit images, PDFs, JSON, XML, and other types directly in
+your browser!
+
+Hovering over a cell also reveals some more controls (such as the delete
+button or the timestamp). Click the title to select a few and do batch
+operations:
+
+If you need some sample data to get started and explore, check out this
+howto create [HBase table
+tutorial](http://gethue.com/hadoop-tutorial-how-to-create-example-tables-in-hbase).
+
+
+### Smart Searchbar
+
+The "Smart Searchbar" is a sophisticated tool that helps you zero-in on
+your data. The smart search supports a number of operations. The most
+basic ones include finding and scanning row keys. Here I am selecting
+two row keys with:
+
+
+    domain.100, domain.200
+
+
+Submitting this query gives me the two rows I was looking for. If I want
+to fetch rows after one of these, I have to do a scan. This is as easy
+as writing a '+' followed by the number of rows you want to fetch.
+
+
+    domain.100, domain.200 +5
+
+
+Fetches domain.100 and domain.200 followed by the next 5 rows. If you're
+ever confused about your results, you can look down below and the query
+bar and also click in to edit your query.
+
+The Smart Search also supports column filtering. On any row, I can
+specify the specific columns or families I want to retrieve. With:
+
+
+    domain.100[column_family:]   
+
+
+I can select a bare family, or mix columns from different families like
+so:
+
+
+    domain.100[family1:, family2:, family3:column_a]
+
+
+Doing this will restrict my results from one row key to the columns I
+specified. If you want to restrict column families only, the same effect
+can be achieved with the filters on the right. Just click to toggle a
+filter.
+
+
+Finally, let's try some more complex column filters. I can query for
+bare columns:
+
+
+    domain.100[column_a]
+
+This will multiply my query over all column families. I can also do
+prefixes and scans:
+
+
+    domain.100[family: prefix* +3]
+
+
+This will fetch me all columns that start with prefix\* limited to 3
+results. Finally, I can filter on range:
+
+
+    domain.100[family: column1 to column100]
+
+
+This will fetch me all columns in 'family:' that are lexicographically
+\>= column1 but <= column100. The first column ('column1') must be a
+valid column, but the second can just be any string for comparison.
+
+The Smart Search also supports prefix filtering on rows. To select a
+prefixed row, simply type the row key followed by a star \*. The prefix
+should be highlighted like any other searchbar keyword. A prefix scan is
+performed exactly like a regular scan, but with a prefixed row.
+
+
+    domain.10* +10
+
+
+Finally, as a new feature, you can also take full advantage of the
+[HBase filtering](denied:about:blank)language, by typing your filter
+string between curly braces. HBase Browser autocompletes your filters
+for you so you don't have to look them up every time. You can apply
+filters to rows or scans.
+
+
+    domain.1000 {ColumnPrefixFilter('100-') AND ColumnCountGetFilter(3)}
+
+
+This doc only covers a few basic features of the Smart Search. You can
+take advantage of the full querying language by referring to the help
+menu when using the app. These include column prefix, bare columns,
+column range, etc. Remember that if you ever need help with the
+searchbar, you can use the help menu that pops up while typing, which
+will suggest next steps to complete your query.
+
+
+
+
+
+## Sqoop 2 Editor
+
+The Sqoop UI enables transfering data from a relational database
+to Hadoop and vice versa. The UI lives uses Apache Sqoop to do this.
+See the [Sqoop Documentation](http://sqoop.apache.org/docs/1.99.2/index.html) for more details on Sqoop.
+
+Installation and Configuration
+------------------------------
+
+The Sqoop UI is one of the applications installed as part of
+Hue. For information about installing and configuring Hue, see the Hue Installation
+manual.
+
+Starting
+--------
+
+Click the **Sqoop** icon
+(![image](images/icon_sqoop_24.png)) in the navigation bar at the top of
+the Hue browser page.
+
+Sqoop Jobs
+----------
+Sqoop UI is oriented around jobs in Apache Sqoop.
+
+### Creating a New Job
+
+1. Click the **New job** button at the top right.
+2. In the Name field, enter a name.
+3. Choose the type of job: import or export.
+   The proceeding form fields will change depending on which type is chosen.
+4. Select a connection, or create one if it does not exist.
+5. Fill in the rest of the fields for the job.
+   For importing, the "Table name", "Storage type", "Output format", and "Output directory" are necessary at a minimum.
+   For exporting, the "Table name" and "Input directory" are necessary at a minimum.
+6. Click **save** to finish.
+
+### Editing a Job
+
+1. In the list of jobs, click on the name of the job.
+2. Edit the desired form fields in the job.
+
+### Copying a Job
+
+1. In the list of jobs, click on the name of the job.
+2. On the left hand side of the job editor, there should be a panel containing actions.
+   Click **Copy**.
+
+### Removing a Job
+
+1. In the list of jobs, click on the name of the job.
+2. On the left hand side of the job editor, there should be a panel containing actions.
+   Click **Delete**.
+
+### Running a Job
+
+There's a status on each of the items in the job list indicating
+the last time a job was ran. The progress of the job should dynamically
+update. There's a progress bar at the bottom of each item on the job list
+as well.
+
+1. In the list of jobs, click on the name of the job.
+2. On the left hand side of the job editor, there should be a panel containing actions.
+   Click **Run**.
+
+### Creating a New Connection
+
+1. Click the **New job** button at the top right.
+2. At the connection field, click the link titled **Add a new connection**.
+3. Fill in the displayed fields.
+4. Click **save** to finish.
+
+### Editing a Connection
+
+1. Click the **New job** button at the top right.
+2. At the connection field, select the connection by name that should be edited.
+3. Click **Edit**.
+4. Edit the any of the fields.
+5. Click **save** to finish.
+
+### Removing a Connection
+
+1. Click the **New job** button at the top right.
+2. At the connection field, select the connection by name that should be deleted.
+3. Click **Delete**.
+
+NOTE: If this does not work, it's like because a job is using that connection.
+      Make sure not jobs are using the connection that will be deleted.
+
+### Filtering Sqoop Jobs
+
+The text field in the top, left corner of the Sqoop Jobs page enables fast filtering
+of sqoop jobs by name.
+
+
+## ZooKeeper Browser
+
+
+The main two features are:
+
+- Listing of the ZooKeeper cluster stats and clients
+- Browsing and edition of the ZNode hierarchy
+
+
+ZooKeeper Browser requires the [ZooKeeper
+REST](https://github.com/apache/zookeeper/tree/trunk/src/contrib/rest)
+service to be running. Here is how to setup this one:
+
+First get and build ZooKeeper:
+
+<pre>
+git clone https://github.com/apache/zookeeper
+cd zookeeper
+ant
+Buildfile: /home/hue/Development/zookeeper/build.xml
+
+init:
+       [mkdir] Created dir: /home/hue/Development/zookeeper/build/classes
+       [mkdir] Created dir: /home/hue/Development/zookeeper/build/lib
+       [mkdir] Created dir: /home/hue/Development/zookeeper/build/package/lib
+       [mkdir] Created dir: /home/hue/Development/zookeeper/build/test/lib
+
+   ...
+</pre>
+
+And start the REST service:
+
+<pre>
+cd src/contrib/rest
+nohup ant run&
+</pre>
+
+If ZooKeeper and the REST service are not on the same machine as Hue, go
+update the [Hue
+settings](https://github.com/cloudera/hue/blob/master/desktop/conf.dist/hue.ini#L581)
+and specify the correct hostnames and ports:
+
+<pre>
+    [zookeeper]
+
+      [[clusters]]
+
+        [[[default]]]
+          # Zookeeper ensemble. Comma separated list of Host/Port.
+          # e.g. localhost:2181,localhost:2182,localhost:2183
+          ## host_ports=localhost:2181
+
+          # The URL of the REST contrib service
+          ## rest_url=http://localhost:9998
+</pre>
+
+
    </div>
 </div>

+ 0 - 110
docs/user-guide/jobbrowser.md

@@ -1,110 +0,0 @@
-
-<link rel="stylesheet" href="docbook.css" type="text/css" media="screen" title="no title" charset="utf-8"></link>
-
-Job Browser
-===========
-
-The Job Browser application lets you to examine the Hadoop MapReduce
-jobs running on your Hadoop cluster. Job Browser presents the job and
-tasks in layers. The top layer is a list of jobs, and you can link to a
-list of that job's tasks. You can then view a task's attempts and the
-properties of each attempt, such as state, start and end time, and
-output size. To troubleshoot failed jobs, you can also view the logs of
-each attempt.
-
-Job Browser Installation and Configuration
-------------------------------------------
-
-Job Browser is one of the applications installed as part of Hue. For
-information about installing and configuring Hue, see the Hue Installation
-manual.
-
-Job Browser can display both MRv1 and MRv2 jobs, but must be configured
-to display one type at a time. 
-
-Starting Job Browser
---------------------
-
-Click the **Job Browser** icon (![image](images/icon_jobbrowser_24.png))
-in the navigation bar at the top of the Hue web page.
-
-If there are no jobs that have been run, the **Welcome to the Job
-Browser** page opens, with links to the Job Designer and Beeswax.
-
-If there are jobs running, then the Job Browser list appears.
-
-Filtering the Job Browser List
-------------------------------
-
--   To filter the jobs by their state (such as **Running** or
-    **Completed**), choose a state from the **Job status** drop-down
-    menu.
--   To filter by a user who ran the jobs, enter the user's name in the
-    **User Name** query box.
--   To filter by job name, enter the name in the **Text** query box.
--   To clear the filters, choose **All States** from the **Job status**
-    drop-down menu and delete any text in the **User Name** and **Text**
-    query boxes.
--   To display retired jobs, check the **Show retired jobs** checkbox.
-    Retired jobs show somewhat limited information – for example,
-    information on maps and reduces and job duration is not available.
-    Jobs are designated as Retired by the JobTracker based on the value
-    of mapred.jobtracker.retirejob.interval. The retired jobs no longer
-    display after the JobTracker is restarted.
-
-Viewing Job Information and Logs
---------------------------------
-
-![image](images/note.jpg) **Note**: At any level you can view the log
-for an object by clicking the ![image](images/log.png) icon in the Logs
-column.
-
-**To view job information for an individual job:**
-
-1.  In the **Job Browser** window, click **View** at the right of the
-    job you want to view. This shows the **Job** page for the job, with
-    the recent tasks associated with the job are displayed in the
-    **Tasks** tab.
-2.  Click the **Metadata** tab to view the metadata for this job.
-3.  Click the **Counters** tab to view the counter metrics for the job.
-
-**To view details about the tasks associated with the job:**
-
-1.  In the **Job** window, click the **View All Tasks** link at the
-    right just above the **Recent Tasks** list. This lists all the tasks
-    associated with the job.
-2.  Click **Attempts** to the right of a task to view the attempts for
-    that task.
-
-**To view information about an individual task:**
-
-1.  In the **Job** window, click the **View** link to the right of the
-    task. The attempts associated with the task are displayed.
-2.  Click the **Metadata** tab to view metadata for this task. The
-    metadata associated with the task is displayed.
-3.  To view the Hadoop counters for a task, click the **Counters** tab.
-    The counters associated with the task are displayed.
-4.  To return to the **Job** window for this job, click the job number
-    in the status panel at the left of the window.
-
-**To view details about a task attempt:**
-
-1.  In the **Job Task** window, click the **View** link to the right of
-    the task attempt. The metadata associated with the attempt is
-    displayed under the **Metadata** tab.
-2.  To view the Hadoop counters for the task attempt, click the
-    **Counters** tab. The counters associated with the attempt are
-    displayed.
-3.  To view the logs associated with the task attempt, click the
-    **Logs** tab. The logs associated with the task attempt are
-    displayed.
-4.  To return to the list of tasks for the current job, click the task
-    number in the status panel at the left of the window.
-
-Viewing Job Output
-------------------
-
-1.  In the **Job Browser** window, click the link in the ID column.
-2.  To view the output of the job, click the link under **OUTPUT** in
-    the panel at the left of the window. This takes you to the job
-    output directory in the **File Browser**.

+ 0 - 333
docs/user-guide/jobdesigner.md

@@ -1,333 +0,0 @@
-
-<link rel="stylesheet" href="docbook.css" type="text/css" media="screen" title="no title" charset="utf-8"></link>
-
-Job Designer
-============
-
-The Job Designer application enables you to create and submit jobs to
-the Hadoop cluster. You can include variables with your jobs to enable
-you and other users to enter values for the variables when they run your
-job. The Job Designer supports the actions supported by
-[Oozie](http://archive.cloudera.com/cdh4/cdh/4/oozie/):
-[MapReduce](/content/cloudera-content/cloudera-docs/HadoopTutorial/CDH4/index.html),
-Streaming, Java, Pig, Hive, Sqoop, Shell, Ssh, DistCp, Fs, and Email.
-
-Job Designer Installation and Configuration
--------------------------------------------
-
-Job Designer is one of the applications installed as part of Hue. For
-information about installing and configuring Hue, see the Hue Installation
-manual..
-
-In order to run DistCp, Streaming, Pig, Sqoop, and Hive jobs, Oozie must
-be configured to use the Oozie ShareLib. See the Oozie Installation manual.
-
-Starting Job Designer
----------------------
-
-Click the **Job Designer** icon (![image](images/icon_jobsub_24.png)) in
-the navigation bar at the top of the Hue web page. The **Job Designs**
-page opens in the browser.
-
-Installing the Example Job Designs
-----------------------------------
-
-![image](images/note.jpg) **Note**: You must be a superuser to perform
-this task.
-
-1.  Click ![image](images/quick_start.png). The Quick Start Wizard
-    opens.
-2.  Click **Step 2: Examples**.
-3.  Click **Job Designer**.
-
-Job Designs
------------
-
-A job design specifies several meta-level properties of a job, including
-the job design name, description, the executable scripts or classes, and
-any parameters for those scripts or classes.
-
-### Filtering Job Designs
-
-You can filter the job designs that appear in the list by owner, name,
-type, and description.
-
-**To filter the Job Designs list:**
-
-1.  In the **Job Designs** window, click **Designs**.
-2.  Enter text in the Filter text box at the top of the **Job Designs**
-    window. When you type in the Filter field, the designs are
-    dynamically filtered to display only those rows containing text that
-    matches the specified substring.
-
-### Creating a Job Design
-
-1.  In the **Job Designs** window, click New Action \> Action, where
-    Action is MapReduce, Streaming, Java, Pig, Hive, Sqoop, Shell, Ssh,
-    DistCp, Fs, or Email.
-2.  In the **Job Design (Action type)** window, specify the common and
-    job type specific information.
-3.  Click **Save** to save the job settings.
-
-### Deleting and Restoring Job Designs
-
-You can move job designs to the trash and later restore or permanently
-delete them.
-
-#### Deleting Job Designs
-
-1.  In a Manager screen, check the checkbox next to one or more job
-    designs.
-2.  Choose one of the following:
-    -   Delete \> Move to trash
-    -   Delete \> Delete forever
-
-#### Restoring Job Designs
-
-1.  In a Manager screen, click ![image](images/trash.png) **Trash**.
-2.  Check the checkbox next to one or more job designs.
-3.  Click Restore.
-
-### Job Design Settings
-
-#### Job Design Common Settings
-
-Most job design types support all the settings listed in the following
-table. For job type specific settings, see:
-[MapReduce](#mapreduce),
-[Streaming](#streaming),
-[Java](#java),
-[Pig](#pig),
-[Hive](#hive),
-[Sqoop](#sqoop),
-[Shell](#shell),
-[Ssh](#ssh),
-[DistCp](#distcp),
-[Fs](#fs), and
-[Email](#email).
-
-All job design settings except Name and Description support the use of
-variables of the form $variable\_name. When you run the job, a dialog
-box will appear to enable you to specify the values of the variables.
-
-<table>
-<tr><td>Name</td><td>Identifies the job and its collection of properties and parameters.</td></tr>
-<tr><td>Description</td><td>A description of the job. The description is displayed in the dialog box
-that appears if you specify variables for the job.</td></tr>
-<tr><td>Advanced</td><td>Advanced settings:<ul><li>Is shared- Indicate whether to share the action with all users.<li>Oozie parameters - parameters to pass to Oozie</td></tr>
-<tr><td>Prepare</td><td>Specifies paths to create or delete before starting the workflow job.</td></tr>
-<tr><td>Params</td>Parameters to pass to a script or command. The parameters are expressed
-using the [JSP 2.0 Specification (JSP.2.3) Expression
-Language](http://jcp.org/aboutJava/communityprocess/final/jsr152/),
-allowing variables, functions, and complex expressions as parameters.<td></td></tr>
-<tr><td>Job Properties</td><td>Job properties. To set a property value, click <b>Add Property</b>.<ol><li>Property name -  a configuration property name. This field provides autocompletion, so you can type the first few characters of a property name and then select the one you want from the drop-down
-    list.<li>Valuethe property value.</td></tr>
-<tr><td>Files</td><td>Files to pass to the job. Equivalent to the Hadoop -files option.</td></tr>
-<tr><td>Archives</td><td>Files to pass to the job. Archives to pass to the job. Equivalent to the Hadoop -archives option.</td></tr></table>
-
-<a id="mapreduce"></a>
-#### MapReduce Job Design
-
-A MapReduce job design consists of MapReduce functions written in Java.
-You can create a MapReduce job design from existing mapper and reducer
-classes without having to write a main Java class. You must specify the
-mapper and reducer classes as well as other MapReduce properties in the
-Job Properties setting.
-
-<table>
-<tr><td>Jar path</td><td>The fully-qualified path to a JAR file containing the classes that
-implement the Mapper and Reducer functions.</td></tr>
-</table>
-
-<a id="streaming"></a>
-#### Streaming Job Design
-
-Hadoop streaming jobs enable you to create MapReduce functions in any
-non-Java language that reads standard Unix input and writes standard
-Unix output. For more information about Hadoop streaming jobs, see
-[Hadoop
-Streaming](http://archive.cloudera.com/cdh/3/hadoop-0.20.2+320/streaming.html).
-
-<table>
-<tr><td>Mapper</td><td>The path to the mapper script or class. If the mapper file is not on the
-machines on the cluster, use the Files option to pass it as a part
-of job submission. Equivalent to the Hadoop -mapper option.</td></tr>
-<tr><td>Reducer</td><td>The path to the reducer script or class. If the reducer file is not on
-the machines on the cluster, use the Files option to pass it as a
-part of job submission. Equivalent to the Hadoop -reducer option.</td></tr>
-</table>
-
-<a id="java"></a>
-#### Java Job Design
-
-A Java job design consists of a main class written in Java.
-
-<table>
-<tr><td>Jar path</td><td>The fully-qualified path to a JAR file containing the main class.</td></tr>
-<tr><td>Main class</td><td>The main class to invoke the program.</td></tr>
-<tr><td>Args</td><td>The arguments to pass to the main class.</td></tr>
-<tr><td>Java opts</td><td>The options to pass to the JVM.</td></tr>
-</table>
-
-<a id="pig"></a>
-#### Pig Job Design
-
-
-A Pig job design consists of a Pig script.
-
-<table>
-<tr><td>Script name</td><td>Script name or path to the Pig script.</td></tr>
-</table>
-
-<a id="hive"></a>
-#### Hive Job Design
-
-A Hive job design consists of a Hive script.
-
-<table>
-<tr><td>Script name</td><td>Script name or path to the Hive script.</td></tr>
-</table>
-
-
-<a id="sqoop"></a>
-#### Sqoop Job Design
-
-A Sqoop job design consists of a Sqoop command.
-
-<table>
-<tr><td>Command</td><td>The Sqoop command.</td></tr>
-</table>
-
-<a id="shell"></a>
-#### Shell Job Design
-
-A Shell job design consists of a shell command.
-
-<table>
-<tr><td>Command</td><td>The shell command.</td></tr>
-<tr><td></td>Capture output<td>Indicate whether to capture the output of the command.</td></tr>
-</table>
-
-<a id="ssh"></a>
-#### Ssh Job Design
-
-A Ssh job design consists of an ssh command.
-
-<table>
-<tr><td>User</td><td>The name of the user to run the command as.</td></tr>
-<tr><td>Host</td><td>The name of the host to run the command on.</td></tr>
-<tr><td>Command</td><td>The ssh command.</td></tr>
-<tr><td></td>Capture output<td>Indicate whether to capture the output of the command.</td></tr>
-</table>
-
-<a id="distcp"></a>
-#### DistCp Job Design
-
-A DistCp job design consists of a DistCp command.
-
-<a id="fs"></a>
-#### Fs Job Design
-
-A Fs job design consists of a command that operates on HDFS.
-
-<table>
-<tr><td>Delete path</td><td>The path to delete. If it is a directory, it deletes recursively all its
-content and then deletes the directory.</td></tr>
-<tr><td></td>Create directory<td>The path of a directory to create.</td></tr>
-<tr><td>Move file</td><td>The source and destination paths to the file to be moved.</td></tr>
-<tr><td>Change permissions</td><td>The path whose permissions are to be changed, the permissions, and an
-indicator of whether to change permission recursively.</td></tr></table>
-
-<a id="email"></a>
-#### Email Job Design
-
-A Email job design consists of an email message.
-
-<table>
-<tr><td>To addresses</td><td>The recipient of the email message.</td></tr>
-<tr><td>CC addresses (optional)</td><td>The cc recipients of the email message.</td></tr>
-<tr><td>Subject</td><td>The subject of the email message.</td></tr>
-<tr><td>Body</td><td>The body of the email message.</td></tr>
-</table>
-
-
-### Submitting a Job Design
-
-![image](images/note.jpg) **Note**:
-
-A job's input files must be uploaded to the cluster before you can
-submit the job.
-
-**To submit a job design:**
-
-1.  In the **Job Designs** window, click **Designs** in the upper left
-    corner. Your jobs and other users' jobs are displayed in the **Job
-    Designs** window.
-2.  Check the checkbox next to the job you want to submit.
-3.  Click the **Submit** button.
-    1.  If the job contains variables, enter the information requested
-        in the dialog box that appears. For example, the sample grep
-        MapReduce design displays a dialog where you specify the output
-        directory.
-    2.  Click **Submit** to submit the job.
-
-After the job is complete, the Job Designer displays the results of the
-job. For information about displaying job results, see [Displaying the
-Results of Submitting a Job](#submitJob).
-
-### Copying, Editing, and Deleting a Job Design
-
-If you want to edit and use a job but you don't own it, you can make a
-copy of it and then edit and use the copied job.
-
-
-**Copy**
-
-1.  In the **Job Designs** window, click **Designs**. The jobs are
-    displayed in the **Job Designs** window.
-2.  Check the checkbox next to the job you want to copy.
-3.  Click the **Copy** button.
-4.  In the **Job Design Editor** window, change the settings and then
-    click **Save** to save the job settings.
-
-**Edit**
-
-1.  In the **Job Designs** window, click **Designs**. The jobs are
-    displayed in the **Job Designs** window.
-2.  Check the checkbox next to the job you want to edit.
-3.  Click the **Edit** button.
-4.  In the **Job Design** window, change the settings and then click
-    **Save** to save the job settings.
-
-Delete
-
-1.  In the **Job Designs** window, click **Designs**. The jobs are
-    displayed in the **Job Designs** window.
-2.  Check the checkbox next to the job you want to delete.
-3.  Click the **Delete** button.
-4.  Click **OK** to confirm the deletion.
-
-<a id="submitJob"></a>
-Displaying Results of Submitting a Job
---------------------------------------
-
-**To display the Job Submission History:**
-
-In the **Job Designs** window, click the **History** tab. The jobs are
-displayed in the **Job Submissions History** listed by Oozie job ID.
-
-**To display Job Details:**
-
-In the **Job Submission History** window, click an Oozie Job ID. The
-results of the job display:
-
--   Actions - a list of actions in the job.
--   Click ![image](images/gear.png) to display the action configuration.
-    In the action configuration for a MapReduce action, click the value
-    of the mapred.output.dir property to display the job output.
--   In the root-node row, click the Id in the External Id column to view
-    the job in the Job Browser.
--   Details - the job details. Click ![image](images/gear.png) to
-    display the Oozie application configuration.
--   Definition - the Oozie application definition.
--   Log - the output log.

+ 0 - 203
docs/user-guide/metastore_manager.md

@@ -1,203 +0,0 @@
-
-<link rel="stylesheet" href="docbook.css" type="text/css" media="screen" title="no title" charset="utf-8"></link>
-
-Metastore Manager
-=================
-
-The Metastore Manager application enables you to manage the databases,
-tables, and partitions of the
-[Hive](http://archive.cloudera.com/cdh4/cdh/4/hive/) metastore shared by
-the ([Beeswax](../beeswax.html) and [Cloudera Impala Query
-UI](../impala.html)) applications. You can use Metastore
-Manager to perform the following operations:
-
--   Databases
-    -   [Select a database](#selectDatabase)
-    -   [Create a database](#createDatabase)
-    -   [Drop databases](#dropDatabase)
-
--   Tables
-    -   [Create tables](#createTables)
-    -   [Browse tables](#browseTables)
-    -   [Import data into a table](#importDataIntoTables)
-    -   [Drop tables](#dropTables)
-    -   [View the location of a table](#viewTableLocation)
-
-Metastore Manager Installation and Configuration
-------------------------------------------------
-
-Metastore Manager is one of the applications installed as part of Hue.
-For information about installing and configuring Hue, see the Hue Installation
-manual.
-
-Starting Metastore Manager
---------------------------
-
-Click the **Metastore Manager** icon
-(![image](images/icon_table_browser_24.png)) in the navigation bar at
-the top of the Hue browser page.
-
-### Installing Sample Tables
-
-![image](images/note.jpg) **Note**: You must be a superuser to perform
-this task.
-
-1.  Click ![image](images/quick_start.png). The Quick Start Wizard
-    opens.
-2.  Click **Step 2: Examples**.
-3.  Click **Beeswax (Hive UI)** or **Cloudera Impala Query UI**.
-
-### Importing Data
-
-If you want to import your own data instead of installing the sample
-tables, follow the procedure in [Creating Tables](#createTables).
-
-<a id="selectDatabase"></a>
-Selecting a Database
---------------------
-
-1.  In the pane on the left, select the database from the DATABASE
-    drop-down list.
-
-<a id="createDatabase"></a>
-Creating a Database
--------------------
-
-1.  Click ![image](images/databases.png).
-2.  Click **Create a new database**.
-    1.  Specify a database name and optional description. Database names
-        are not case-sensitive. Click **Next**.
-    2.  Do one of the following:
-        -   Keep the default location in the Hive warehouse folder.
-        -   Specify an external location within HDFS:
-            1.  Uncheck the **Location** checkbox.
-            2.  In the External location field, type a path to a folder
-                on HDFS or click ![image](images/browse.png) to browse
-                to a folder and click **Select this folder**.
-
-    3.  Click the **Create Database** button.
-    
-<a id="selectDatabase"></a>
-Dropping Databases
-------------------
-
-1.  Click ![image](images/databases.png).
-2.  In the list of databases, check the checkbox next to one or more
-    databases.
-3.  Click the ![image](images/trash.png) Drop button.
-4.  Confirm whether you want to delete the databases.
-
-<a id="createTables"></a>
-Creating Tables
----------------
-
-Although you can create tables by executing the appropriate Hive HQL DDL
-query commands, it is easier to create a table using the Metastore
-Manager table creation wizard.
-
-There are two ways to create a table: from a file or manually. If you
-create a table from a file, the format of the data in the file will
-determine some of the properties of the table, such as the record and
-file formats. The data from the file you specify is imported
-automatically upon table creation. When you create a file manually, you
-specify all the properties of the table, and then execute the resulting
-query to actually create the table. You then import data into the table
-as an additional step.
-
-**From a File**
-
-1.  In the ACTIONS pane in the Metastore Manager window, click **Create
-    a new table from a file**. The table creation wizard starts.
-2.  Follow the instructions in the wizard to create the table. The basic
-    steps are:
-    -   Choose your input file. The input file you specify must exist.
-        Note that you can choose to have Beeswax create the table
-        definition only based on the import file you select, without
-        actually importing data from that file.
-    -   Specify the column delimiter.
-    -   Define your columns, providing a name and selecting the type.
-
-3.  Click **Create Table** to create the table. The new table's metadata
-    displays on the right side of the **Table Metadata** window. At this
-    point, you can view the metadata or a sample of the data in the
-    table. From the ACTIONS pane you can import new data into the table,
-    browse the table, drop it, or go to the File Browser to see the
-    location of the data.
-
-**Manually**
-
-1.  In the ACTIONS pane in the Metastore Manager window, click **Create
-    a new table manually**. The table creation wizard starts.
-2.  Follow the instructions in the wizard to create the table. The basic
-    steps are:
-    -   Name the table.
-    -   Choose the record format.
-    -   Configure record serialization by specifying delimiters for
-        columns, collections, and map keys.
-    -   Choose the file format.
-    -   Specify the location for your table's data.
-    -   Specify the columns, providing a name and selecting the type for
-        each column.
-    -   Specify partition columns, providing a name and selecting the
-        type for each column.
-
-3.  Click **Create table**. The Table Metadata window displays.
-
-<a id="browseTables"></a>
-Browsing Tables
----------------
-
-**To browse table data:**
-
-In the Table List window, check the checkbox next to a table name and
-click **Browse Data**. The table's data displays in the Query Results
-window.
-
-**To browse table metadata:**
-
-Do one of the following:
-
--   In the Table List window, click a table name.
--   Check the checkbox next to a table name and click **View**.
-
--   The table's metadata displays in the **Columns** tab. You can view
-    the table data by selecting the **Sample** tab.
--   If the table is partitioned, you can view the partition columns by
-    clicking the **Partition Columns** tab and display the partitions by
-    clicking **Show Partitions(n)**, where n is the number of partitions
-    in the ACTIONS pane on the left.
-
-<a id="importDataIntoTables"></a>
-Importing Data into a Table
----------------------------
-
-When importing data, you can choose to append or overwrite the table's
-data with data from a file.
-
-1.  In the Table List window, click the table name. The Table Metadata
-    window displays.
-2.  In the ACTIONS pane, click **Import Data**.
-3.  For **Path**, enter the path to the file that contains the data you
-    want to import.
-4.  Check **Overwrite existing data** to replace the data in the
-    selected table with the imported data. Leave unchecked to append to
-    the table.
-5.  Click **Submit**.
-
-<a id="dropTables"></a>
-Dropping Tables
----------------
-
-1.  In the Table List window, click the table name. The Table Metadata
-    window displays.
-2.  In the ACTIONS pane, click **Drop Table**.
-3.  Click **Yes** to confirm the deletion.
-
-<a id="viewTableLocation"></a>
-Viewing a Table's Location
---------------------------
-
-1.  In the Table List window, click the table name. The Table Metadata
-    window displays.
-2.  Click **View File Location**. The file location of the selected
-    table displays in its directory in the File Browser window.

+ 0 - 585
docs/user-guide/oozie.md

@@ -1,585 +0,0 @@
-
-<link rel="stylesheet" href="docbook.css" type="text/css" media="screen" title="no title" charset="utf-8"></link>
-
-Oozie Editor and Dashboard
-==========================
-
-The Oozie Editor/Dashboard application allows you to define Oozie
-workflow, coordinator, and bundle applications, run workflow,
-coordinator, and bundle jobs, and view the status of jobs. For
-information about Oozie, see [Oozie
-Documentation](http://archive.cloudera.com/cdh4/cdh/4/oozie/).
-
-A workflow application is a collection of actions arranged in a directed
-acyclic graph (DAG). It includes two types of nodes:
-
--   Control flow - start, end, fork, join, decision, and kill
--   Action - [MapReduce](../jobdesigner.html#mapreduce),
-[Streaming](../jobdesigner.html#streaming),
-[Java](../jobdesigner.html#java),
-[Pig](../jobdesigner.html#pig),
-[Hive](../jobdesigner.html#hive),
-[Sqoop](../jobdesigner.html#sqoop),
-[Shell](../jobdesigner.html#shell),
-[Ssh](../jobdesigner.html#ssh),
-[DistCp](../jobdesigner.html#distcp),
-[Fs](../jobdesigner.html#fs), and
-[Email](../jobdesigner.html#email).
-    In order to run DistCp, Streaming, Pig, Sqoop, and Hive jobs, Oozie
-    must be configured to use the Oozie ShareLib. See the Oozie Installation
-   manual.
-
-A coordinator application allows you to define and execute recurrent and
-interdependent workflow jobs. The coordinator application defines the
-conditions under which the execution of workflows can occur.
-
-A bundle application allows you to batch a set of coordinator
-applications.
-
-Oozie Editor/Dashboard Installation and Configuration
------------------------------------------------------
-
-Oozie Editor/Dashboard is one of the applications installed as part of
-Hue. For information about installing and configuring Hue, see the Hue Installation
-manual.
-
-Starting Oozie Editor/Dashboard
--------------------------------
-
-Click the **Oozie Editor/Dashboard** icon
-(![image](images/icon_oozie_24.png)) in the navigation bar at the top of
-the Hue browser page. **Oozie Editor/Dashboard** opens with the
-following screens:
-
--   [Dashboard](#dashboard) - shows the running and completed workflow,
-    coordinator, and bundle jobs and information about Oozie
-    instrumentation and configuration. The screen is selected and opened
-    to the Workflows page.
--   [Workflow Manager](#workflowManager) - shows available workflows and
-    allows you to create and import workflows.
--   [Coordinator Manager](#coordinatorManager) - shows available coordinators and
-    allows you to create coordinators.
--   [Bundle Manager](#bundleManager) - shows available bundles and
-    allows you to create bundles.
-
-Installing Oozie Editor/Dashboard Examples
-------------------------------------------
-
-![image](images/note.jpg) **Note**: You must be a superuser to perform
-this task.
-
-1.  Click ![image](images/quick_start.png). The Quick Start Wizard
-    opens.
-2.  Click **Step 2: Examples**.
-3.  Click **Oozie Editor/Dashboard**.
-
-Filtering Lists in Oozie Editor/Dashboard
------------------------------------------
-
-Many screens contain lists. When you type in the Filter field on
-screens, the lists are dynamically filtered to display only those rows
-containing text that matches the specified substring.
-
-Permissions in Oozie Editor/Dashboard
--------------------------------------
-
-In the Dashboard workflows, coordinators, and bundles can only be
-viewed, submitted, and modified by their owner or a superuser.
-
-Editor permissions for performing actions on workflows, coordinators,
-and bundles are summarized in the following table:
-
-<table>
-<th><td>Action</td><td>Superuser or Owner</td><td>All</td></th>
-<tr><td>View</td><td>Y.</td><td>Only if "Is shared" is set</td></tr>
-<tr><td>Submit</td><td>Y.</td><td>Only if "Is shared" is set</td></tr>
-<tr><td>Modify</td><td>Y.</td><td>N</td></tr>
-</table>
-
-
-Deleting and Restoring Workflows, Coordinators, and Bundles
------------------------------------------------------------
-
-You can move workflows, coordinators, and bundles to the trash and later
-restore or permanently delete them.
-
-### Deleting Workflows, Coordinators, and Bundles
-
-1.  In a Manager screen, check the checkbox next to one or more
-    workflows, coordinators or bundles.
-2.  Choose one of the following:
-    -   Delete \> Move to trash
-    -   Delete \> Delete forever
-
-### Restoring Workflows, Coordinators, and Bundles
-
-1.  In a Manager screen, click ![image](images/trash.png) **Trash**.
-2.  Check the checkbox next to one or more workflows, coordinators or
-    bundles.
-3.  Click Restore.
-
-
-<a id="dashboard"></a>
-Dashboard
----------
-
-The Dashboard shows a summary of the running and completed workflow,
-coordinator, and bundle jobs.
-
-You can view jobs for a period up to the last 30 days.
-
-You can filter the list by date (1, 7, 15, or 30 days) or status
-(Succeeded, Running, or Killed). The date and status buttons are
-toggles.
-
-
-### Workflows
-
-Click the **Workflows** tab to view the running and completed workflow
-jobs for the filters you have specified.
-
-Click a workflow row in the Running or Completed table to view detailed
-information about that workflow job.
-
-In the left pane contains a link to the workflow and the variable values
-specified.![image](images/workflow.jpg)
-
-For the selected job, the following information is available in the
-right area.
-
--   **Graph** tab shows the workflow DAG.
--   **Actions** tab shows you details about the actions that make up the
-    workflow.
-    -   Click the **Id** link to see additional details about the
-        action.
-    -   Click the **External Id** link to view the job in the Job
-        Browser.
-
--   **Details** tab shows job statistics including start and end times.
--   **Configuration** tab shows selected job configuration settings.
--   **Logs** tab shows log output generated by the workflow job.
--   **Definition** tab shows the Oozie workflow definition, as it
-    appears in the workflow.xml file (also linked under the application
-    path properties in the **Details** tab and the **Configuration**
-    tab).
-
-For each action in the workflow you can:
-
--   Click the ![image](images/eye.png) icon to view the action screen,
-    which contains:
-    -   **Details** tab shows job statistics including start and end
-        times.
-    -   **Configuration** tab shows the action configuration settings.
-    -   **Child Jobs** tab lists jobs generated by the action.
-
--   Click the ![image](images/log.png) icon to view the log in the Job
-    Browser.
-
-
-### Coordinators
-
-Click the **Coordinators** tab to view the running and completed
-coordinator jobs for the filters you have specified.
-
-For the selected job, the following information is available.
-
--   The **Calendar** tab shows the timestamp of the job. Click the
-    timestamp to open the workflow DAG.
--   The **Actions** tab shows you details about the actions that make up
-    the coordinator.
-    -   Click the **Id** link to see additional details about the
-        action.
-    -   Click the **External Id** link to view the job in the Job
-        Browser.
-
--   The **Configuration** tab shows selected job configuration settings.
--   The **Logs** tab shows log output generated by the coordinator.
--   The **Definition** tab shows the Oozie coordinator definition, as it
-    appears in the coordinator.xml file (also linked under the
-    oozie.coord.application.path property in the **Configuration** tab).
-
-### Bundles
-
-Click the **Bundles** tab to view the running and completed bundle jobs
-for the filters you have specified.
-
-### Oozie
-
-The Oozie tab provides subtabs that give you access to Oozie
-instrumentation and configuration settings.
-
-#### Instrumentation
-
-For information on the instrumentation metrics supported by Oozie, see
-[Oozie
-Monitoring](http://oozie.apache.org/docs/3.3.0/AG_Monitoring.html).
-
-#### Configuration
-
-For information on the configuration properties supported by Oozie, see
-[Oozie
-Configuration](http://oozie.apache.org/docs/3.3.0/AG_Install.html#Oozie_Configuration).
-
-<a id="workflowManager"></a>
-Workflow Manager
-----------------
-
-In Workflow Manager you create Oozie workflows and submit them for
-execution.
-
-Click the **Workflows** tab to open the Workflow Manager.
-
-Each row shows a workflow: its name, description, timestamp of its last
-modification. It also shows:
-
--   **Steps** - the number of steps in the workflow execution path. This
-    is the number of execution steps between the start and end of the
-    workflow. This will not necessarily be the same as the number of
-    actions in the workflow, if there are control flow nodes in the
-    control path.
--   **Status** - who can run the workflow. **shared** means users other
-    than the owner can access the workflow. **personal** means only the
-    owner can modify or submit the workflow. The default is personal.
--   **Owner** - the user that created the workflow.
-
-In Workflow Editor you edit workflows that include MapReduce, Streaming,
-Java, Pig, Hive, Sqoop, Shell, Ssh, DistCp, Fs, Email, Sub-workflow, and
-Generic actions. You can configure these actions in the Workflow Editor,
-or you can import job designs from Job Designer to be used as actions in
-your workflow. For information about defining workflows, see the
-[Workflow
-Specification](http://archive.cloudera.com/cdh4/cdh/4/oozie/WorkflowFunctionalSpec.html).
-
-### Opening a Workflow
-
-To open a workflow, in Workflow Manager, click the workflow. Proceed
-with [Editing a Workflow](#editingWorkflow).
-
-### Creating a Workflow
-
-1.  Click the **Create** button at the top right.
-2.  In the Name field, type a name.
-3.  Check the Is shared checkbox to allow all users to access the
-    workflow.
-4.  Click **advanced** to specify the deployment directory or a job.xml
-    file.
-5.  Click **Save**. The Workflow Editor opens. Proceed with [Editing a
-    Workflow](#editingWorkflow).
-
-### Importing a Workflow
-
-1.  Click the **Import** button at the top right.
-2.  In the Name field, type a name.
-3.  In the **Local workflow.xml file** field, click **Choose File** and
-    select a workflow file.
-4.  Click **advanced** to specify whether the workflow is shared, the
-    deployment directory, or a job.xml file.
-5.  Click **Save**. The Workflow Editor opens. Proceed with [Editing a
-    Workflow](#editingWorkflow).
-
-### Submitting a Workflow
-
-To submit a workflow for execution, do one of the following:
-
--   In the Workflow Manager, click the radio button next to the
-    workflow, and click the **Submit** button.
--   In the Workflow Editor, click the **Submit** button.
-
-The workflow job is submitted and the Dashboard displays the workflow
-job.
-
-To view the output of the job, click ![image](images/log.png) **View the
-logs**.
-
-#### Suspending a Running Job
-
-In the pane on the left, click the **Suspend** button.
-
-1.  Verify that you want to suspend the job.
-
-#### Resuming a Suspended Job
-
-In the pane on the left, click the **Resume** button.
-
-1.  Verify that you want to resume the job.
-
-#### Rerunning a Workflow
-
-In the pane on the left, click the **Rerun** button.
-
-1.  Check the checkboxes next to the actions to rerun.
-2.  Specify required variables.
-3.  Click **Submit**.
-
-### Scheduling a Workflow
-
-To schedule a workflow for recurring execution, do one of the following:
-
--   In the Workflow Manager, check the checkbox next to the workflow and
-    click the **Schedule** button.
--   In the Workflow Editor, click the **Schedule** button.
-
-A coordinator is created and opened in the Coordinator Editor. Proceed
-with [Editing a Coordinator](#editingCoordinator).
-
-<a id="editingWorkflow"></a>
-### Editing a Workflow
-
-In the Workflow Editor you can easily perform operations on Oozie action
-and control nodes.
-
-#### Action Nodes
-
-The Workflow Editor supports dragging and dropping action nodes. As you
-move the action over other actions and forks, highlights indicate active
-areas. If there are actions in the workflow, the active areas are the
-actions themselves and the areas above and below the actions. If you
-drop an action on an existing action, a fork and join is added to the
-workflow.
-
--   Add actions to the workflow by clicking an action
-    ![image](images/action.png) button and drop the action on the
-    workflow. The Edit Node screen displays.
-    1.  Set the action properties and click **Done**. Each action in a
-        workflow must have a unique name.
-
--   Copy an action by clicking the ![image](images/copy.png) **Copy**
-    button.
-
-1.  The action is opened in the Edit Node screen.
-2.  Edit the action properties and click **Done**. The action is added
-    to the end of the workflow.
-
--   Delete an action by clicking the ![image](images/trash.png) button.
--   Edit an action by clicking the ![image](images/edit.png) button.
--   Change the position of an action by left-clicking and dragging an
-    action to a new location.
-
-#### Control Nodes
-
--   Create a fork and join by dropping an action on top of another
-    action.
--   Remove a fork and join by dragging a forked action and dropping it
-    above the fork.
--   Convert a fork to a decision by clicking the
-    ![image](images/convert.png) button.
--   To edit a decision:
-    1.  Click the ![image](images/edit.png) button.
-    2.  Fill in the predicates that determine which action to perform
-        and select the default action from the drop-down list.
-    3.  Click **Done**.
-
-### Uploading Workflow Files
-
-In the Workflow Editor, click the **Upload** button.
-
-The workspace of the workflow is opened in the File Browser application.
-Follow the procedure in [Uploading
-Files](../filebrowser.html#uploadingFiles) to upload the files. You must
-put JAR files in a lib directory in the workspace.
-
-### Editing Workflow Properties
-
-1.  In the Workflow Editor, click the link under the Name or Description
-    fields in the left pane.
-2.  To share the workflow with all users, check the **Is shared**
-    checkbox.
-3.  To set advanced execution options, click **advanced** and edit the
-    deployment directory, add parameters and job properties, or specify
-    a job.xml file.
-4.  Click **Save**.
-
-### Displaying the History of a Workflow
-
-1.  Click the **Dashboard** tab.
-2.  Click the **Workflows** tab.
-3.  Click a workflow.
-4.  Click the **Actions** tab.
-
-<a id="coordinatorManager"></a>
-Coordinator Manager
--------------------
-
-In Coordinator Manager you create Oozie coordinator applications and
-submit them for execution.
-
-Click the **Coordinators** tab to open the Coordinator Manager.
-
-Each row shows a coordinator: its name, description, timestamp of its
-last modification. It also shows:
-
--   **Workflow** - the workflow that will be run by the coordinator.
--   **Frequency** - how often the workflow referenced by the coordinator
-    will be run.
--   **Status** - who can run the coordinator. **shared** means users
-    other than the owner can access the workflow. **personal** means
-    only the owner can modify or submit the workflow. The default is
-    personal.
--   **Owner** - the user that created the coordinator.
-
-In Coordinator Editor, you edit coordinators and the datasets required
-by the coordinators. For information about defining coordinators and
-datasets, see the [Coordinator
-Specification](http://archive.cloudera.com/cdh4/cdh/4/oozie/CoordinatorFunctionalSpec.html).
-
-### Opening a Coordinator
-
-To open a coordinator, in Coordinator Manager, click the coordinator.
-Proceed with [Editing a Coordinator](#editingCoordinator).
-
-### Creating a Coordinator
-
-To create a coordinator, in Coordinator Manager:
-
-1.  Click the **Create** button at the top right. The Coordinator wizard
-    opens. Proceed with [Editing a Coordinator](#editingCoordinator).
-
-### Submitting a Coordinator
-
-To submit a coordinator for execution, check the checkbox next to the
-coordinator and click the **Submit** button.
-
-<a id="editingCoordinator"></a>
-### Editing a Coordinator
-
-In the Coordinator Editor you specify coordinator properties and the
-datasets on which the workflow scheduled by the coordinator will operate
-by stepping through screens in a wizard. You can also advance to
-particular steps and revisit steps by clicking the Step "tabs" above the
-screens. The following instructions walk you through the wizard.
-
-1.  Type a name, select the workflow, check the **Is shared checkbox**
-    to share the job, and click **Next**. If the Coordinator Editor was
-    opened after scheduling a workflow, the workflow will be set.
-2.  Select how many times the coordinator will run for each specified
-    unit, the start and end times of the coordinator, the timezone of
-    the start and end times, and click **Next**. Times must be expressed
-    as UTC times. For example, to run at 10 pm PST, specify a start time
-    of 6 am UTC of the following day (+8 hours) and set the Timezone
-    field to America/Los\_Angeles.
-3.  Click **Add** to select an input dataset and click **Next**. If no
-    datasets exist, follow the procedure in [Creating a
-    Dataset](#creatingDataset).
-4.  Click **Add** to select an output dataset. Click **Save
-    coordinator** or click **Next** to specify advanced settings.
-5.  To share the coordinator with all users, check the **Is****shared**
-    checkbox.
-6.  Fill in parameters to pass to Oozie, properties that determine how
-    long a coordinator will wait before timing out, how many
-    coordinators can run and wait concurrently, and the coordinator
-    execution policy.
-7.  Click **Save coordinator**.
-
-<a id="creatingDataset"></a>
-### Creating a Dataset
-
-1.  In the Coordinator Editor, do one of the following:
-    -   Click **here** in the Inputs or Outputs pane at the top of the
-        editor.
-    -   In the pane at the left, click the **Create new** link. Proceed
-        with [Editing a Dataset](#editingDataset).
-
-### Displaying Datasets
-
-1.  In the Coordinator Editor, click **Show existing** in pane at the
-    left.
-2.  To edit a dataset, click the dataset name in the Existing datasets
-    table. Proceed with [Editing a Dataset](#editingDataset).
-
-<a id="editingDataset"></a>
-### Editing a Dataset
-
-1.  Type a name for the dataset.
-2.  In the Start and Frequency fields, specify when and how often the
-    dataset will be available.
-3.  In the URI field, specify a URI template for the location of the
-    dataset. To construct URIs and URI paths containing dates and
-    timestamps, you can specify the variables
-    ${YEAR},${MONTH},${DAY},${HOUR},${MINUTE}. For example:
-    hdfs://foo:9000/usr/app/stats/${YEAR}/${MONTH}/data.
-4.  In the Instance field, click a button to choose a default, single,
-    or range of data instances. For example, if frequency==DAY, a window
-    of the last rolling 5 days (not including today) would be expressed
-    as start: -5 and end: -1. Check the advanced checkbox to display a
-    field where you can specify a coordinator [EL
-    function](http://archive.cloudera.com/cdh4/cdh/4/oozie/CoordinatorFunctionalSpec.html).
-5.  Specify the timezone of the start date.
-6.  In the Done flag field, specify the flag that identifies when input
-    datasets are no longer ready.
-
-### Displaying the History of a Coordinator
-
-1.  Click the **Dashboard** tab.
-2.  Click the **Coordinators** tab.
-3.  Click a coordinator.
-4.  Click the **Actions** tab.
-
-<a id="bundleManager"></a>
-Bundle Manager
---------------
-
-In Bundle Manager you create Oozie bundle applications and submit them
-for execution.
-
-Click the **Bundle** tab to open the Bundle Manager.
-
-Each row shows a bundle: its name, description, timestamp of its last
-modification. It also shows:
-
--   **Coordinators** - the coordinators that will be run by the bundle.
--   **Kick off** - the UTC time when the coordinators referenced by the
-    bundle will be started.
--   **Status** - who can run the bundle. **shared** means users other
-    than the owner can access the workflow. **personal** means only the
-    owner can modify or submit the workflow. The default is personal.
--   **Owner** - the user that created the bundle.
-
-For information about defining bundles, see the [Bundle
-Specification](http://archive.cloudera.com/cdh4/cdh/4/oozie/BundleFunctionalSpec.html).
-
-### Opening a Bundle
-
-To open a bundle, in Bundle Manager, click the bundle. Proceed with
-[Editing a Bundle](#editingBundle).
-
-### Creating a Bundle
-
-1.  Click the **Create** button at the top right.
-2.  In the Name field, type a name.
-3.  In the Kick off time field, choose a kick off time.
-4.  Check the Is shared checkbox to allow all users to access the
-    workflow.
-5.  Click **Save**. The Bundle Editor opens. Proceed with [Editing a
-    Bundle](#editingBundle).
-
-### Submitting a Bundle
-
-To submit a bundle for execution, check the checkbox next to the bundle
-and click the **Submit** button.
-
-<a id="editingBundle"></a>
-### Editing a Bundle
-
-In the Bundle Editor, you specify properties by stepping through screens
-in a wizard. You can also advance to particular steps and revisit steps
-by clicking the Step "tabs" above the screens. The following
-instructions walk you through the wizard.
-
-1.  Click **Add** to select a coordinator that the bundle will kick off.
-2.  Choose the kick off time. The time must be expressed as a UTC time.
-    For example, to run at 10 pm PST, specify a start time of 6 am UTC
-    of the following day (+8 hours).
-3.  To share the bundle with all users, check the **Is shared**
-    checkbox.
-4.  Click **Next** to specify advanced settings or click **Save
-    bundle**.
-5.  Fill in parameters to pass to Oozie.
-6.  Click **Save bundle**.
-
-Displaying the History of a Bundle
-----------------------------------
-
-1.  Click the **Dashboard** tab.
-2.  Click the **Bundles** tab.
-3.  Click a bundle.
-4.  Click the **Actions** tab.

+ 0 - 71
docs/user-guide/pig.md

@@ -1,71 +0,0 @@
-
-<link rel="stylesheet" href="docbook.css" type="text/css" media="screen" title="no title" charset="utf-8"></link>
-
-Pig Editor
-==========
-
-The Pig Editor application allows you to define Pig scripts, run
-scripts, and view the status of jobs. For information about Pig, see
-[Pig Documentation](http://archive.cloudera.com/cdh4/cdh/4/pig/).
-
-Pig Editor Installation and Configuration
------------------------------------------
-
-Pig Editor is one of the applications installed as part of Hue. For
-information about installing and configuring Hue, see the Hue Installation manual.
-
-Pig Editor assumes an existing Pig installation. The Hue installation
-instructions include the configuration necessary for Pig Editor to
-access Pig.
-
-Starting Pig Editor
--------------------
-
-Click the **Pig Editor** icon (![image](images/icon_pig_24.png)) in the
-navigation bar at the top of the Hue browser page. The Pig Editor opens
-with three tabs:
-
--   Editor - editor where you can create, edit, run, save, copy, and
-    delete scripts and edit script properties.
--   Scripts - script manager where you can create, open, run, copy, and
-    delete scripts.
--   Dashboard - dashboard where you can view running and completed
-    scripts and view the log of a job.
-
-Pig Scripts
------------
-
-### Installing the Example Scripts
-
-![image](images/note.jpg) **Note**: You must be a superuser to perform
-this task.
-
-1.  Click ![image](images/quick_start.png). The Quick Start Wizard
-    opens.
-2.  Click **Step 2: Examples**.
-3.  Click **Pig Editor**.
-
-### Creating a Script
-
-1.  In either the Editor or Scripts screen, click New script. Edit the
-    script as desired.
-2.  Click Edit Properties. In the Script name field, type a name for the
-    script.
-3.  Click Save.
-
-### Opening a Script
-
-1.  Click the Scripts tab.
-2.  In the list of scripts, click a script.
-
-### Running a Script
-
-1.  Do one of the following:
-    -   In the Editor screen, click the Run button.
-    -   In the Scripts screen, check the checkbox next to a script and
-        click the Run button.
-
-### Viewing the Result of Running a Script
-
-1.  Click the Dashboard tab.
-2.  Click a job.

+ 0 - 124
docs/user-guide/search.md

@@ -1,124 +0,0 @@
-
-<link rel="stylesheet" href="docbook.css" type="text/css" media="screen" title="no title" charset="utf-8"></link>
-
-# Solr Search
-
-
-The Solr Search application, which is based on  [Apache Solr](http://lucene.apache.org/solr/), allows you to perform keyword searches across Hadoop data. A wizard lets you style the result snippets, specify facets to group the results, sort the results, and highlight result fields.
-
-## Solr Search Installation and Configuration
-
-Solr Search is one of the applications installed as part of Hue. For information about installing and configuring Hue, see the Hue Installation
-manual.
-
-## Starting  Solr Search
-
-Click the ** Solr Search** icon
-(![image](images/icon_search_24.png)) in the navigation bar at the top of
-the Hue browser page. **Solr Search** opens to the  [Collection Manager](#collectionManager). If there are no collections, the [Import Collections and Cores](#importCollection) dialog displays.
-
-<a id="collectionManager"></a>
-## Collection Manager
-
-In Collection Manager you import, copy, and delete [collections](http://wiki.apache.org/solr/SolrCloud#A_little_about_SolrCores_and_Collections).
-
-### Displaying the Collection Manager
-When you start Solr Search, the Collection Manager displays. You navigate to the Collection Manager by clicking **Collection manager** in the Search page or the Template Editor.
-
-###  Filtering Collections
-
-When you type in the Filter field, the list of collections is dynamically filtered to display only those rows
-containing text that matches the specified substring.
-
-<a id="importCollection"></a>
-### Importing Collections 
-
-1. If there are existing collections, click the ![image](images/import.png) **Import** button at the top right. The Import Collections and Cores dialog displays.
-1. Check the checkboxes next to the collections to import.
-1. Click **Import Selected**. The collection is added to the Collection Manager.
-
-### Editing Collection Properties
-1. In the Collection Manager, click a collection.
-1. In the **COLLECTION** area on the left, click **Properties**.
-1. Edit a property and click **Save**.
-
-
-### Searching a Collection
-
-1. In the Collection Manager, click **Search page** or click **Search it** in the Collection area on the left. The Search page displays.
-1. Select a collection from the **Search in** drop-down list.
-1. Type a search string in the **Search...** text box.
-1. Press **Enter** or click the ![image](images/eyeglass.png)  icon.
-
--  If you have defined [facets](#facets), click a facet to display only those results in the group defined by the facet.
--  If you have defined [sorting fields](#sorting), select from the **Sort by** drop-down list to sort the results.
--  Click ![image](images/clear.png) to clear the search string.
-
-## Styling Search Results
-
-Do one of the following:
-
-- In the Collection Manager, click a collection.
-- In the Search page, select a collection from the **Search in** drop-down list and click ** Customize this collection**.  The Template Editor displays.
-
-### Template Editor
-
-The Template Editor provides four features:
-
-- [Snippet editor](#snippetEditor) - Specify the layout of the search result snippet, which fields appear in the snippet, and style the results.
-- [Facet editor](#facetEditor) - Define buckets in which to group results.
-- [Sort editor](#sortEditor) - Specify on which fields and order the results are sorted. 
-- [Highlighting editor](#highlightingEditor) - Enable highlighting of search fields. 
-
-<a id="snippetEditor"></a>
-#### Snippet Editor
-
-1. In the Snippet Editor, click a tab to choose the method for editing the search snippet fields and styling:
-1. - **Visual editor** - Click ![image](images/layoutChooser.png) to choose an overall layout for the snippet.
-1. - - Select the fields and functions from the drop-down lists on the right and click ![image](images/add.png). 
-1. - - Select fields, right-click, and select **Cut** and **Paste** to place the fields on the canvas. 
-1. - - Select fields and apply styling using the buttons on top.
-1. - **Source** - 
-1. - - Select the data fields and functions from the drop-down lists on the right.
-1. - - Specify layout and styling using HTML tags.
-1. - **Preview** - Preview the snippet.
-1. - **Advanced** - Specify styles for CSS classes specified in the Source tab.
-1. Click **Save**.
-
-<a id="facetEditor"></a>
-#### Facet Editor
-
-By default, faceting  search result fields is disabled. Click **Enabled** to enable faceting.
-
-1. In the Template Editor, click **2. Facets**. You can move between the facet tabs by clicking each **Step** tab, or by clicking **Back** and **Next**.
-1. In the General tab, specify 
-1. - **Limit** - the maximum number of values for each facet.
-1. - **Mincount** - the minimum number of search results that fall into a group for the facet to display on the Search page.
-1. In the Field, Range, and Date Facet tabs,  specify the facet properties and click ![image](images/add.png) **Add**.
-1. In the Facets Order tab, drag and drop the facet to specify the order in they appear in the Search page.
-1. Click **Save**. When you display the Search page, the facets display on the left.
-
-<a id="sortEditor"></a>
-#### Sorting Editor 
-
-By default, sorting on search result fields is disabled. Click **Enabled** to enable sorting.
-
-1. In the Template Editor, click **3. Sorting**. 
-1. In the Field drop-down, select a field. Optionally specify a label for the field. 
-1. The default order is ascending. Click the arrows to change the order.
-1. Click  ![image](images/add.png) **Add**.
-1. Click **Save**. When you display search results, the results are sorted by the fields in the order that  they appear left to right. 
-
-<a id="highlightingEditor"></a>
-#### Highlighting Editor 
-
-By default,highlighting search result fields is disabled. Click **Enabled** to enable highlighting.
-
-1. In the Template Editor, click **3. Highlighting**. 
-1. Select the fields to be highlighted.
-1. Click **Save**. When you display search results, the selected fields are displayed with the style of the **em** class defined in the Advanced tab of the [Snippet editor](#snippetEditor). 
-
-
-
-
-

+ 0 - 63
docs/user-guide/shell.md

@@ -1,63 +0,0 @@
-
-<link rel="stylesheet" href="docbook.css" type="text/css" media="screen" title="no title" charset="utf-8"></link>
-
-Hue Shell
-=========
-
-The Hue Shell application provides access to the Pig, HBase, and Sqoop 2
-command-line shells. The Shell application is designed to have the same
-look and feel as a Unix terminal. In addition to the shells configured
-by default, it is possible to include almost any process that exposes a
-command-line interface as an option in this Hue application.
-
-Hue Shell Installation and Configuration
-----------------------------------------
-
-Hue Shell is one of the applications installed as part of Hue. For
-information about installing and configuring Hue, see Hue Installation
-in
-[http://www.cloudera.com/content/cloudera-content/cloudera-docs/CDH4/latest/CDH4-Installation-Guide/CDH4-Installation-Guide.html](http://www.cloudera.com/content/cloudera-content/cloudera-docs/CDH4/latest/CDH4-Installation-Guide/CDH4-Installation-Guide.html).
-
-### Unix User Accounts
-
-To properly isolate subprocesses so as to guarantee security, each Hue
-user who is using the Shell subprocess must have a Unix user account.
-The link between Hue users and Unix user accounts is the username, and
-so every Hue user who wants to use the Shell application must have a
-Unix user account with the same name on the system that runs the Hue
-Server. See Unix User Accounts in
-[http://www.cloudera.com/content/cloudera-content/cloudera-docs/CDH4/latest/CDH4-Installation-Guide/CDH4-Installation-Guide.html](http://www.cloudera.com/content/cloudera-content/cloudera-docs/CDH4/latest/CDH4-Installation-Guide/CDH4-Installation-Guide.html)
-for instructions.
-
-Starting Hue Shell
-------------------
-
-1.  Click the **Shell** icon (![image](images/icon_shell_24.png)) in the
-    navigation bar at the top of the Hue web page. (To start a second
-    instance of the Shell application, right-click the link and select
-    **Open link in new tab**.) The **Shell** window opens in the Hue web
-    page.
-2.  Click any of the tabs at the top of the Shell window to open a
-    subprocess shell of that type.
-    ![image](images/note.jpg) **Note**: If a button is disabled, the
-    program is not on the system path. Ask your Hue administrator to fix
-    this problem.
-3.  After opening a subprocess, click anywhere in the body of the Shell
-    application window to focus the command line. (Tab-completion is not
-    supported.)
-4.  To end a process, type exit or quit depending on the type of
-    subprocess you have opened.
-    ![image](images/note.jpg) **Note**: If you close your browser, the
-    underlying shell process remains running for the amount of time
-    specified by your Hue administrator. After this time the process is
-    killed.
-
-Viewing Documentation for the Shells
-------------------------------------
-
-For information about using each of the default shells, see the
-documentation on the following sites:
-
--   [Pig](http://archive.cloudera.com/cdh4/cdh/4/pig/)
--   [HBase](http://archive.cloudera.com/cdh4/cdh/4/hbase/)
--   [Sqoop 2](http://archive.cloudera.com/cdh4/cdh/4/sqoop2/)

+ 0 - 96
docs/user-guide/sqoop.md

@@ -1,96 +0,0 @@
-
-<link rel="stylesheet" href="docbook.css" type="text/css" media="screen" title="no title" charset="utf-8"></link>
-
-Sqoop UI
-========
-
-The Sqoop UI enables transfering data from a relational database
-to Hadoop and vice versa. The UI lives uses Apache Sqoop to do this.
-See the [Sqoop Documentation](http://sqoop.apache.org/docs/1.99.2/index.html) for more details on Sqoop.
-
-Installation and Configuration
-------------------------------
-
-The Sqoop UI is one of the applications installed as part of
-Hue. For information about installing and configuring Hue, see the Hue Installation
-manual.
-
-Starting
---------
-
-Click the **Sqoop** icon
-(![image](images/icon_sqoop_24.png)) in the navigation bar at the top of
-the Hue browser page.
-
-Sqoop Jobs
-----------
-Sqoop UI is oriented around jobs in Apache Sqoop.
-
-### Creating a New Job
-
-1. Click the **New job** button at the top right.
-2. In the Name field, enter a name.
-3. Choose the type of job: import or export.
-   The proceeding form fields will change depending on which type is chosen.
-4. Select a connection, or create one if it does not exist.
-5. Fill in the rest of the fields for the job.
-   For importing, the "Table name", "Storage type", "Output format", and "Output directory" are necessary at a minimum.
-   For exporting, the "Table name" and "Input directory" are necessary at a minimum.
-6. Click **save** to finish.
-
-### Editing a Job
-
-1. In the list of jobs, click on the name of the job.
-2. Edit the desired form fields in the job.
-
-### Copying a Job
-
-1. In the list of jobs, click on the name of the job.
-2. On the left hand side of the job editor, there should be a panel containing actions.
-   Click **Copy**.
-
-### Removing a Job
-
-1. In the list of jobs, click on the name of the job.
-2. On the left hand side of the job editor, there should be a panel containing actions.
-   Click **Delete**.
-
-### Running a Job
-
-There's a status on each of the items in the job list indicating
-the last time a job was ran. The progress of the job should dynamically
-update. There's a progress bar at the bottom of each item on the job list
-as well.
-
-1. In the list of jobs, click on the name of the job.
-2. On the left hand side of the job editor, there should be a panel containing actions.
-   Click **Run**.
-
-### Creating a New Connection
-
-1. Click the **New job** button at the top right.
-2. At the connection field, click the link titled **Add a new connection**.
-3. Fill in the displayed fields.
-4. Click **save** to finish.
-
-### Editing a Connection
-
-1. Click the **New job** button at the top right.
-2. At the connection field, select the connection by name that should be edited.
-3. Click **Edit**.
-4. Edit the any of the fields.
-5. Click **save** to finish.
-
-### Removing a Connection
-
-1. Click the **New job** button at the top right.
-2. At the connection field, select the connection by name that should be deleted.
-3. Click **Delete**.
-
-NOTE: If this does not work, it's like because a job is using that connection.
-      Make sure not jobs are using the connection that will be deleted.
-
-### Filtering Sqoop Jobs
-
-The text field in the top, left corner of the Sqoop Jobs page enables fast filtering
-of sqoop jobs by name.

+ 0 - 59
docs/user-guide/zookeeper.md

@@ -1,59 +0,0 @@
-
-<link rel="stylesheet" href="docbook.css" type="text/css" media="screen" title="no title" charset="utf-8"></link>
-
-ZooKeeper Browser
-=================
-
-The main two features are:
-
-- Listing of the ZooKeeper cluster stats and clients
-- Browsing and edition of the ZNode hierarchy
-
-
-ZooKeeper Browser requires the [ZooKeeper
-REST](https://github.com/apache/zookeeper/tree/trunk/src/contrib/rest)
-service to be running. Here is how to setup this one:
-
-First get and build ZooKeeper:
-
-<pre>
-git clone https://github.com/apache/zookeeper
-cd zookeeper
-ant
-Buildfile: /home/hue/Development/zookeeper/build.xml
-
-init:
-       [mkdir] Created dir: /home/hue/Development/zookeeper/build/classes
-       [mkdir] Created dir: /home/hue/Development/zookeeper/build/lib
-       [mkdir] Created dir: /home/hue/Development/zookeeper/build/package/lib
-       [mkdir] Created dir: /home/hue/Development/zookeeper/build/test/lib
-
-   ...
-</pre>
-
-And start the REST service:
-
-<pre>
-cd src/contrib/rest
-nohup ant run&
-</pre>
-
-If ZooKeeper and the REST service are not on the same machine as Hue, go
-update the [Hue
-settings](https://github.com/cloudera/hue/blob/master/desktop/conf.dist/hue.ini#L581)
-and specify the correct hostnames and ports:
-
-<pre>
-    [zookeeper]
-
-      [[clusters]]
-
-        [[[default]]]
-          # Zookeeper ensemble. Comma separated list of Host/Port.
-          # e.g. localhost:2181,localhost:2182,localhost:2183
-          ## host_ports=localhost:2181
-
-          # The URL of the REST contrib service
-          ## rest_url=http://localhost:9998
-</pre>
-