
Let's make lava_job_submitter.py cleaner with only parsing and retry mechanism capabilities. Moved out from the submitter script: 1. proxy functions - moved to lava.utils.lava_proxy.py 2. LAVAJob class definition - moved to lava.utils.lava_job.py - added structural logging capabilities into LAVAJob - Implemented properties for job_id, is_finished, and status, with corresponding setter methods that update the log dictionary. - Added new methods show, get_lava_time, and refresh_log for improved log handling and data retrieval. Signed-off-by: Guilherme Gallo <guilherme.gallo@collabora.com> Part-of: <https://gitlab.freedesktop.org/mesa/mesa/-/merge_requests/22500>
45 lines
1.3 KiB
Python
45 lines
1.3 KiB
Python
import time
|
|
import traceback
|
|
import urllib
|
|
import urllib.parse
|
|
import xmlrpc
|
|
import xmlrpc.client
|
|
|
|
import lavacli
|
|
|
|
from .log_follower import fatal_err, print_log
|
|
|
|
|
|
def setup_lava_proxy():
|
|
config = lavacli.load_config("default")
|
|
uri, usr, tok = (config.get(key) for key in ("uri", "username", "token"))
|
|
uri_obj = urllib.parse.urlparse(uri)
|
|
uri_str = f"{uri_obj.scheme}://{usr}:{tok}@{uri_obj.netloc}{uri_obj.path}"
|
|
transport = lavacli.RequestsTransport(
|
|
uri_obj.scheme,
|
|
config.get("proxy"),
|
|
config.get("timeout", 120.0),
|
|
config.get("verify_ssl_cert", True),
|
|
)
|
|
proxy = xmlrpc.client.ServerProxy(uri_str, allow_none=True, transport=transport)
|
|
|
|
print_log(f'Proxy for {config["uri"]} created.')
|
|
|
|
return proxy
|
|
|
|
|
|
def call_proxy(fn, *args):
|
|
retries = 60
|
|
for n in range(1, retries + 1):
|
|
try:
|
|
return fn(*args)
|
|
except xmlrpc.client.ProtocolError as err:
|
|
if n == retries:
|
|
traceback.print_exc()
|
|
fatal_err(f"A protocol error occurred (Err {err.errcode} {err.errmsg})")
|
|
else:
|
|
time.sleep(15)
|
|
except xmlrpc.client.Fault as err:
|
|
traceback.print_exc()
|
|
fatal_err(f"FATAL: Fault: {err.faultString} (code: {err.faultCode})", err)
|