Source code for airflow.providers.amazon.aws.operators.batch
## Licensed to the Apache Software Foundation (ASF) under one# or more contributor license agreements. See the NOTICE file# distributed with this work for additional information# regarding copyright ownership. The ASF licenses this file# to you under the Apache License, Version 2.0 (the# "License"); you may not use this file except in compliance# with the License. You may obtain a copy of the License at## http://www.apache.org/licenses/LICENSE-2.0## Unless required by applicable law or agreed to in writing,# software distributed under the License is distributed on an# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY# KIND, either express or implied. See the License for the# specific language governing permissions and limitations# under the License.#"""An Airflow operator for AWS Batch services.. seealso:: - http://boto3.readthedocs.io/en/latest/guide/configuration.html - http://boto3.readthedocs.io/en/latest/reference/services/batch.html - https://docs.aws.amazon.com/batch/latest/APIReference/Welcome.html"""fromtypingimportAny,Dict,Optionalfromairflow.exceptionsimportAirflowExceptionfromairflow.modelsimportBaseOperatorfromairflow.providers.amazon.aws.hooks.batch_clientimportAwsBatchClientHook
[docs]classAwsBatchOperator(BaseOperator):""" Execute a job on AWS Batch :param job_name: the name for the job that will run on AWS Batch (templated) :type job_name: str :param job_definition: the job definition name on AWS Batch :type job_definition: str :param job_queue: the queue name on AWS Batch :type job_queue: str :param overrides: the `containerOverrides` parameter for boto3 (templated) :type overrides: Optional[dict] :param array_properties: the `arrayProperties` parameter for boto3 :type array_properties: Optional[dict] :param parameters: the `parameters` for boto3 (templated) :type parameters: Optional[dict] :param job_id: the job ID, usually unknown (None) until the submit_job operation gets the jobId defined by AWS Batch :type job_id: Optional[str] :param waiters: an :py:class:`.AwsBatchWaiters` object (see note below); if None, polling is used with max_retries and status_retries. :type waiters: Optional[AwsBatchWaiters] :param max_retries: exponential back-off retries, 4200 = 48 hours; polling is only used when waiters is None :type max_retries: int :param status_retries: number of HTTP retries to get job status, 10; polling is only used when waiters is None :type status_retries: int :param aws_conn_id: connection id of AWS credentials / region name. If None, credential boto3 strategy will be used. :type aws_conn_id: str :param region_name: region name to use in AWS Hook. Override the region_name in connection (if provided) :type region_name: str :param tags: collection of tags to apply to the AWS Batch job submission if None, no tags are submitted :type tags: dict .. note:: Any custom waiters must return a waiter for these calls: .. code-block:: python waiter = waiters.get_waiter("JobExists") waiter = waiters.get_waiter("JobRunning") waiter = waiters.get_waiter("JobComplete") """
[docs]defexecute(self,context:Dict):""" Submit and monitor an AWS Batch job :raises: AirflowException """self.submit_job(context)self.monitor_job(context)
[docs]defon_kill(self):response=self.hook.client.terminate_job(jobId=self.job_id,reason="Task killed by the user")self.log.info("AWS Batch job (%s) terminated: %s",self.job_id,response)
[docs]defmonitor_job(self,context:Dict):""" Monitor an AWS Batch job monitor_job can raise an exception or an AirflowTaskTimeout can be raised if execution_timeout is given while creating the task. These exceptions should be handled in taskinstance.py instead of here like it was previously done :raises: AirflowException """ifnotself.job_id:raiseAirflowException('AWS Batch job - job_id was not found')ifself.waiters:self.waiters.wait_for_job(self.job_id)else:self.hook.wait_for_job(self.job_id)self.hook.check_job_success(self.job_id)self.log.info("AWS Batch job (%s) succeeded",self.job_id)