瀏覽代碼

HUE-7248 [adls] Avoid core libs import dependencies on apps

Romain Rigaux 8 年之前
父節點
當前提交
63cbe94

+ 3 - 2
apps/filebrowser/src/filebrowser/settings.py

@@ -21,8 +21,9 @@ REQUIRES_HADOOP = False
 ICON = "filebrowser/art/icon_filebrowser_48.png"
 MENU_INDEX = 20
 
-PERMISSION_ACTION_S3 = "s3_access"
-PERMISSION_ACTION_ADLS = "adls_access"
+from aws.s3.s3fs import PERMISSION_ACTION_S3
+from azure.adls.webhdfs import PERMISSION_ACTION_ADLS
+
 
 PERMISSION_ACTIONS = (
   (PERMISSION_ACTION_S3, "Access to S3 from filebrowser and filepicker."),

+ 6 - 0
desktop/core/src/desktop/lib/fs/proxyfs.py

@@ -16,9 +16,15 @@
 
 from __future__ import absolute_import
 
+import logging
+
 from urlparse import urlparse
 from django.contrib.auth.models import User
 
+
+LOG = logging.getLogger(__name__)
+
+
 class ProxyFS(object):
 
   def __init__(self, filesystems_dict, default_scheme):

+ 1 - 3
desktop/libs/aws/src/aws/s3/s3fs.py

@@ -35,11 +35,9 @@ from aws.conf import get_default_region, get_locations
 from aws.s3 import normpath, s3file, translate_s3_error, S3A_ROOT
 from aws.s3.s3stat import S3Stat
 
-from filebrowser.settings import PERMISSION_ACTION_S3
-
 
 DEFAULT_READ_SIZE = 1024 * 1024  # 1MB
-
+PERMISSION_ACTION_S3 = "s3_access"
 BUCKET_NAME_PATTERN = re.compile("^((?:(?:[a-zA-Z0-9]|[a-zA-Z0-9][a-zA-Z0-9_\-]*[a-zA-Z0-9])\.)*(?:[A-Za-z0-9]|[A-Za-z0-9][A-Za-z0-9_\-]*[A-Za-z0-9]))$")
 
 LOG = logging.getLogger(__name__)

+ 7 - 2
desktop/libs/azure/src/azure/adls/webhdfs.py

@@ -21,20 +21,25 @@ Interfaces for ADLS via HttpFs/WebHDFS
 import logging
 import threading
 
-from filebrowser.settings import PERMISSION_ACTION_ADLS
+from urlparse import urlparse
+
 from hadoop.fs.webhdfs import WebHdfs as HadoopWebHdfs
 from hadoop.fs.exceptions import WebHdfsException
 from hadoop.hdfs_site import get_umask_mode
+
 from desktop.lib.rest import http_client, resource
 from azure.conf import get_default_adls_url, get_default_adls_fs
-from urlparse import urlparse
+
 
 LOG = logging.getLogger(__name__)
 
 #Azure has a 30MB block limit on upload.
 UPLOAD_CHUCK_SIZE = 30 * 1000 * 1000
+PERMISSION_ACTION_ADLS = "adls_access"
+
 
 class WebHdfs(HadoopWebHdfs):
+
   def __init__(self, url,
                fs_defaultfs,
                logical_name=None,