Эх сурвалжийг харах

HUE-5202 Use DEFAULT_USER.get() instead of hardcoding hue user

krish 9 жил өмнө
parent
commit
c403d215f7

+ 8 - 0
apps/filebrowser/src/filebrowser/conf.py

@@ -44,3 +44,11 @@ SHOW_UPLOAD_BUTTON = Config(
   help=_("whether to show the upload button in hdfs file browser."),
   help=_("whether to show the upload button in hdfs file browser."),
   type=coerce_bool,
   type=coerce_bool,
   default=True)
   default=True)
+
+
+ENABLE_EXTRACT_UPLOADED_ARCHIVE = Config(
+  key="enable_extract_uploaded_archive",
+  help=_("Flag to enable the extraction of a uploaded archive in HDFS."),
+  type=bool,
+  default=False
+)

+ 1 - 1
apps/filebrowser/src/filebrowser/views.py

@@ -59,6 +59,7 @@ from hadoop.fs.hadoopfs import Hdfs
 from hadoop.fs.exceptions import WebHdfsException
 from hadoop.fs.exceptions import WebHdfsException
 from hadoop.fs.fsutils import do_overwrite_save
 from hadoop.fs.fsutils import do_overwrite_save
 
 
+from filebrowser.conf import ENABLE_EXTRACT_UPLOADED_ARCHIVE
 from filebrowser.conf import MAX_SNAPPY_DECOMPRESSION_SIZE
 from filebrowser.conf import MAX_SNAPPY_DECOMPRESSION_SIZE
 from filebrowser.conf import SHOW_DOWNLOAD_BUTTON
 from filebrowser.conf import SHOW_DOWNLOAD_BUTTON
 from filebrowser.conf import SHOW_UPLOAD_BUTTON
 from filebrowser.conf import SHOW_UPLOAD_BUTTON
@@ -68,7 +69,6 @@ from filebrowser.lib import xxd
 from filebrowser.forms import RenameForm, UploadFileForm, UploadArchiveForm, MkDirForm, EditorForm, TouchForm,\
 from filebrowser.forms import RenameForm, UploadFileForm, UploadArchiveForm, MkDirForm, EditorForm, TouchForm,\
                               RenameFormSet, RmTreeFormSet, ChmodFormSet, ChownFormSet, CopyFormSet, RestoreFormSet,\
                               RenameFormSet, RmTreeFormSet, ChmodFormSet, ChownFormSet, CopyFormSet, RestoreFormSet,\
                               TrashPurgeForm
                               TrashPurgeForm
-from notebook.conf import ENABLE_EXTRACT_UPLOADED_ARCHIVE
 
 
 
 
 DEFAULT_CHUNK_SIZE_BYTES = 1024 * 4 # 4KB
 DEFAULT_CHUNK_SIZE_BYTES = 1024 * 4 # 4KB

+ 1 - 1
apps/filebrowser/src/filebrowser/views_test.py

@@ -35,8 +35,8 @@ from desktop.lib.django_test_util import make_logged_in_client
 from desktop.lib.test_utils import grant_access, add_to_group
 from desktop.lib.test_utils import grant_access, add_to_group
 from hadoop import pseudo_hdfs4
 from hadoop import pseudo_hdfs4
 from hadoop.conf import UPLOAD_CHUNK_SIZE
 from hadoop.conf import UPLOAD_CHUNK_SIZE
+from filebrowser.conf import ENABLE_EXTRACT_UPLOADED_ARCHIVE
 from filebrowser.views import location_to_url
 from filebrowser.views import location_to_url
-from notebook.conf import ENABLE_EXTRACT_UPLOADED_ARCHIVE
 
 
 from conf import MAX_SNAPPY_DECOMPRESSION_SIZE
 from conf import MAX_SNAPPY_DECOMPRESSION_SIZE
 from lib.rwx import expand_mode
 from lib.rwx import expand_mode

+ 9 - 8
desktop/core/src/desktop/lib/tasks/extract_archive/extract_utils.py

@@ -15,8 +15,9 @@
 # See the License for the specific language governing permissions and
 # See the License for the specific language governing permissions and
 # limitations under the License.
 # limitations under the License.
 
 
-from desktop.conf import DEFAULT_USER
 from notebook.connectors.base import Notebook
 from notebook.connectors.base import Notebook
+from desktop.conf import DEFAULT_USER
+
 
 
 def extract_archive_in_hdfs(request, upload_path, file_name):
 def extract_archive_in_hdfs(request, upload_path, file_name):
 
 
@@ -27,16 +28,16 @@ def extract_archive_in_hdfs(request, upload_path, file_name):
       shell_command='extract_archive_in_hdfs.sh',
       shell_command='extract_archive_in_hdfs.sh',
       arguments=[{'value': '-u=' + upload_path}, {'value': '-f=' + file_name}],
       arguments=[{'value': '-u=' + upload_path}, {'value': '-f=' + file_name}],
       archives=[],
       archives=[],
-      files=[{'value': '/user/hue/common/extract_archive_in_hdfs.sh'}, {"value": upload_path + '/' + file_name}],
+      files=[{'value': '/user/' + DEFAULT_USER.get() + '/common/extract_archive_in_hdfs.sh'}, {"value": upload_path + '/' + file_name}],
       env_var=[{'value': 'HADOOP_USER_NAME=${wf:user()}'}])
       env_var=[{'value': 'HADOOP_USER_NAME=${wf:user()}'}])
   return shell_notebook.execute(request, batch=True)
   return shell_notebook.execute(request, batch=True)
 
 
 def _upload_extract_archive_script_to_hdfs(fs):
 def _upload_extract_archive_script_to_hdfs(fs):
-  if not fs.exists('/user/hue/common/'):
-    fs.do_as_user(DEFAULT_USER.get(), fs.mkdir, '/user/hue/common/')
-    fs.do_as_user(DEFAULT_USER.get(), fs.chmod, '/user/hue/common/', 0755)
+  if not fs.exists('/user/' + DEFAULT_USER.get() + '/common/'):
+    fs.do_as_user(DEFAULT_USER.get(), fs.mkdir, '/user/' + DEFAULT_USER.get() + '/common/')
+    fs.do_as_user(DEFAULT_USER.get(), fs.chmod, '/user/' + DEFAULT_USER.get() + '/common/', 0755)
 
 
-  if not fs.do_as_user(DEFAULT_USER.get(), fs.exists, '/user/hue/common/extract_archive_in_hdfs.sh'):
+  if not fs.do_as_user(DEFAULT_USER.get(), fs.exists, '/user/' + DEFAULT_USER.get() + '/common/extract_archive_in_hdfs.sh'):
     fs.do_as_user(DEFAULT_USER.get(), fs.copyFromLocal, 'desktop/core/src/desktop/lib/extract_archive_in_hdfs.sh',
     fs.do_as_user(DEFAULT_USER.get(), fs.copyFromLocal, 'desktop/core/src/desktop/lib/extract_archive_in_hdfs.sh',
-                          '/user/hue/common/extract_archive_in_hdfs.sh')
-    fs.do_as_user(DEFAULT_USER.get(), fs.chmod, '/user/hue/common/', 0755)
+                          '/user/' + DEFAULT_USER.get() + '/common/extract_archive_in_hdfs.sh')
+    fs.do_as_user(DEFAULT_USER.get(), fs.chmod, '/user/' + DEFAULT_USER.get() + '/common/', 0755)

+ 0 - 7
desktop/libs/notebook/src/notebook/conf.py

@@ -102,13 +102,6 @@ ENABLE_QUERY_SCHEDULING = Config(
   default=False
   default=False
 )
 )
 
 
-ENABLE_EXTRACT_UPLOADED_ARCHIVE = Config(
-  key="enable_extract_uploaded_archive",
-  help=_t("Flag to enable the extraction of a uploaded archive in HDFS."),
-  type=bool,
-  default=False
-)
-
 ENABLE_BATCH_EXECUTE = Config(
 ENABLE_BATCH_EXECUTE = Config(
   key="enable_batch_execute",
   key="enable_batch_execute",
   help=_t("Flag to enable the bulk submission of queries as a background task through Oozie."),
   help=_t("Flag to enable the bulk submission of queries as a background task through Oozie."),