xqute.schedulers.gbatch_scheduler
xqute.schedulers.gbatch_scheduler
GbatchScheduler— Scheduler for Google Cloud Batch</>
xqute.schedulers.gbatch_scheduler.GbatchScheduler(*args, project, location, mount=None, service_account=None, network=None, subnetwork=None, no_external_ip_address=None, machine_type=None, provisioning_model=None, image_uri=None, entrypoint=None, commands=None, runnables=None, **kwargs)
Scheduler for Google Cloud Batch
You can pass extra configuration parameters to the constructor
that will be used in the job configuration file.
For example, you can pass taskGroups to specify the task groups
and their specifications.
For using containers, it is a little bit tricky to specify the commands.
When no entrypoint is specified, the commands should be a list
with the first element being the interpreter (e.g. /bin/bash)
and the second element being the path to the wrapped job script.
If the entrypoint is specified, we can use the {lang} and {script}
placeholders in the commands list, where {lang} will be replaced
with the interpreter (e.g. /bin/bash) and {script} will be replaced
with the path to the wrapped job script.
With entrypoint specified and no {script} placeholder, the joined command
will be the interpreter followed by the path to the wrapped job script will be
appended to the commands list.
**kwargs— Other arguments for the schedulerproject(str) — GCP project IDlocation(str) — GCP location (e.g. us-central1)mount(Union, optional) — GCS path to mount (e.g. gs://my-bucket:/mnt/my-bucket)You can pass a list of mounts. You can also use named mount likeNAME=gs://bucket/dirthen it will be mounted to/mnt/disks/NAMEin the container. You can use environment variableNAMEin your job scripts to refer to the mounted path.service_account(str | none, optional) — GCP service account email (e.g. test-account@example.com)network(str | none, optional) — GCP network (e.g. default-network)subnetwork(str | none, optional) — GCP subnetwork (e.g. regions/us-central1/subnetworks/default)no_external_ip_address(bool | none, optional) — Whether to disable external IP addressmachine_type(str | none, optional) — GCP machine type (e.g. e2-standard-4)provisioning_model(str | none, optional) — GCP provisioning model (e.g. SPOT)image_uri(str | none, optional) — Container image URI (e.g. ubuntu-2004-lts)entrypoint(str, optional) — Container entrypoint (e.g. /bin/bash)commands(Union, optional) — The command list to run in the container.There are three ways to specify the commands: 1. If no entrypoint is specified, the final command will be [commands, wrapped_script], where the entrypoint is the wrapper script interpreter that is determined byJOBCMD_WRAPPER_LANG(e.g. /bin/bash), commands is the list you provided, and wrapped_script is the path to the wrapped job script. 2. You can specify something like "-c", then the final command will be ["-c", "wrapper_script_interpreter, wrapper_script"] 3. You can use the placeholders{lang}and{script}in the commands list, where{lang}will be replaced with the interpreter (e.g. /bin/bash) and{script}will be replaced with the path to the wrapped job script. For example, you can specify ["{lang} {script}"] and the final command will be ["wrapper_interpreter, wrapper_script"]runnables(Optional, optional) — Additional runnables to run before or after the main job.Each runnable should be a dictionary that follows the GCP Batch API specification. You can also specify an "order" key in the dictionary to control the execution order of the runnables. Runnables with negative order will be executed before the main job, and those with non-negative order will be executed after the main job. The main job runnable will always be executed in the order it is defined in the list.
jobcmd_wrapper_init— The init script for the job command wrapperjobcmd_wrapper_init(str) — The init script for the job command wrapper</>name— The name of the scheduler
check_all_done(jobs,polling_counter)(bool) — Check if all jobs are done (full polling with hooks)</>count_running_jobs(jobs)(int) — Count currently running/active jobs (lightweight check)</>create_job(index,cmd,envs)(Job) — Create a job</>job_fails_before_running(job)(bool) — Check if a job fails before running.</>job_is_running(job)(bool) — Check if a job is really running</>job_is_submitted_or_running(job)(bool) — Check if a job is already submitted or running</>jobcmd_end(job)(str) — The job command end</>jobcmd_init(job)(str) — The job command init</>jobcmd_prep(job)(str) — The job command preparation</>jobcmd_shebang(job)(str) — The shebang of the wrapper script</>kill_job(job)— Kill a job</>kill_job_and_update_status(job)— Kill a job and update its status</>kill_running_jobs(jobs)— Try to kill all running jobs</>retry_job(job)— Retry a job</>submit_job(job)(str) — Submit a job</>submit_job_and_update_status(job)— Submit and update the status</>transition_job_status(job,new_status,old_status,flush,rc,error_msg,is_killed)— Centralized status transition handler</>wrap_job_script(job)(str) — Wrap the job script</>wrapped_job_script(job)(SpecPath) — Get the wrapped job script</>
create_job(index, cmd, envs=None)
Create a job
index(int) — The index of the jobcmd(Union) — The command of the job
The job
submit_job_and_update_status(job)
Submit and update the status
- Check if the job is already submitted or running
- If not, run the hook
- If the hook is not cancelled, clean the job
- Submit the job, raising an exception if it fails
- If the job is submitted successfully, update the status
- If the job fails to submit, update the status and write stderr to the job file
job(Job) — The job
transition_job_status(job, new_status, old_status=None, flush=True, rc=None, error_msg=None, is_killed=False)
Centralized status transition handler
Handles all aspects of job status transitions:
- - Status change logging
- - Hook lifecycle management (ensuring on_job_started is called)
- - Appropriate hook calls based on new status
- - RC file updates
- - Error message appending to stderr
- - JID file cleanup for terminal states
- - Pipeline halt on errors if configured
Note that this method will not flush status changes to disk (job.status_file).You need to call job.set_status() separately if needed.
job(Job) — The job to transitionnew_status(int) — The new status to transition toold_status(int | none, optional) — The previous status (if known).If None, will use job._statusflush(bool, optional) — Whether to flush the status to diskrc(str | none, optional) — Optional return code to write to rc_fileerror_msg(str | none, optional) — Optional error message to append to stderr_fileis_killed(bool, optional) — Whether this is a killed job (uses on_job_killed hook)
kill_job_and_update_status(job)
Kill a job and update its status
job(Job) — The job
count_running_jobs(jobs)
Count currently running/active jobs (lightweight check)
This is optimized for the producer to check if new jobs can be submitted. It only counts jobs without refreshing status or calling hooks.
jobs(List) — The list of jobs
Number of jobs currently in active states
check_all_done(jobs, polling_counter)
Check if all jobs are done (full polling with hooks)
This does complete status refresh and calls all lifecycle hooks. Used by the main polling loop to track job completion.
jobs(List) — The list of jobspolling_counter(int) — The polling counter for hook calls
True if all jobs are done, False otherwise
kill_running_jobs(jobs)
Try to kill all running jobs
jobs(List) — The list of jobs
job_is_submitted_or_running(job)
Check if a job is already submitted or running
job(Job) — The job
True if yes otherwise False.
jobcmd_shebang(job) → str
The shebang of the wrapper script
jobcmd_prep(job) → str
The job command preparation
jobcmd_end(job) → str
The job command end
wrap_job_script(job)
wrapped_job_script(job)
job_fails_before_running(job)
Check if a job fails before running.
For some schedulers, the job might fail before running (after submission). For example, the job might fail to allocate resources. In such a case, the wrapped script might not be executed, and the job status will not be updated (stays in SUBMITTED). We need to check such jobs and mark them as FAILED.
For the instant scheduler, for example, the local scheduler, the failure will be immediately reported when submitting the job, so we don't need to check such jobs.
job(Job) — The job to check
True if the job fails before running, otherwise False.
jobcmd_init(job) → str
The job command init