Browse Source

HUE-8758 [spark] Properly rename job server to LivyClient

Romain 5 năm trước cách đây
mục cha
commit
e4fd81769b

+ 1 - 1
apps/jobbrowser/src/jobbrowser/apis/livy_api.py

@@ -19,7 +19,7 @@ import logging
 
 from django.utils.translation import ugettext as _
 
-from spark.job_server_api import get_api
+from spark.livy_client import get_api
 
 from jobbrowser.apis.base_api import Api
 

+ 1 - 1
apps/spark/src/spark/conf.py

@@ -100,7 +100,7 @@ def get_livy_server_url():
   return url
 
 def get_spark_status(user):
-  from spark.job_server_api import get_api
+  from spark.livy_client import get_api
   status = None
 
   try:

+ 3 - 3
apps/spark/src/spark/job_server_api.py → apps/spark/src/spark/livy_client.py

@@ -45,14 +45,14 @@ def get_api(user):
     API_CACHE_LOCK.acquire()
     try:
       if API_CACHE is None:
-        API_CACHE = JobServerApi(get_livy_server_url())
+        API_CACHE = LivyClient(get_livy_server_url())
     finally:
       API_CACHE_LOCK.release()
   API_CACHE.setuser(user)
   return API_CACHE
 
 
-class JobServerApi(object):
+class LivyClient(object):
 
   def __init__(self, livy_url):
     self._url = posixpath.join(livy_url)
@@ -71,7 +71,7 @@ class JobServerApi(object):
     self._client.set_verify(SSL_CERT_CA_VERIFY.get())
 
   def __str__(self):
-    return "JobServerApi at %s" % (self._url,)
+    return "LivyClient at %s" % (self._url,)
 
   @property
   def url(self):

+ 1 - 1
desktop/libs/notebook/src/notebook/connectors/spark_batch.py

@@ -22,7 +22,7 @@ LOG = logging.getLogger(__name__)
 
 
 try:
-  from spark.job_server_api import get_api as get_spark_api
+  from spark.livy_client import get_api as get_spark_api
 except ImportError as e:
   LOG.exception('Spark is not enabled')
 

+ 1 - 1
desktop/libs/notebook/src/notebook/connectors/spark_shell.py

@@ -38,7 +38,7 @@ LOG = logging.getLogger(__name__)
 
 try:
   from spark.conf import LIVY_SERVER_SESSION_KIND
-  from spark.job_server_api import get_api as get_spark_api
+  from spark.livy_client import get_api as get_spark_api
 except ImportError as e:
   LOG.exception('Spark is not enabled')
 

+ 0 - 0
desktop/libs/notebook/src/notebook/connectors/spark_tests.py → desktop/libs/notebook/src/notebook/connectors/spark_shell_tests.py