[AIRFLOW-6313] Unify example or doc dag owner (#6864)
We already change Airflow DAG default owner to 'airflow' in https://github.com/apache/airflow/pull/4151 but some of our example DAGs and docs are still using owner = 'Airflow', This patch to unify them
This commit is contained in:
Родитель
6cf276be67
Коммит
65ef8f3f81
|
@ -25,7 +25,7 @@ from airflow import DAG
|
|||
from airflow.contrib.operators.azure_container_instances_operator import AzureContainerInstancesOperator
|
||||
|
||||
default_args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'depends_on_past': False,
|
||||
'start_date': datetime(2018, 11, 1),
|
||||
'email': ['airflow@example.com'],
|
||||
|
|
|
@ -33,7 +33,7 @@ from airflow.contrib.sensors.azure_cosmos_sensor import AzureCosmosDocumentSenso
|
|||
from airflow.utils import dates
|
||||
|
||||
default_args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'depends_on_past': False,
|
||||
'start_date': dates.days_ago(2),
|
||||
'email': ['airflow@example.com'],
|
||||
|
|
|
@ -37,7 +37,7 @@ from airflow import DAG
|
|||
from airflow.contrib.operators.databricks_operator import DatabricksSubmitRunOperator
|
||||
|
||||
default_args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'email': ['airflow@example.com'],
|
||||
'depends_on_past': False,
|
||||
'start_date': airflow.utils.dates.days_ago(2)
|
||||
|
|
|
@ -26,7 +26,7 @@ from airflow import DAG
|
|||
from airflow.contrib.operators.dingding_operator import DingdingOperator
|
||||
|
||||
args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'retries': 3,
|
||||
'start_date': airflow.utils.dates.days_ago(2)
|
||||
}
|
||||
|
|
|
@ -27,7 +27,7 @@ from airflow.contrib.operators.emr_create_job_flow_operator import EmrCreateJobF
|
|||
from airflow.contrib.sensors.emr_job_flow_sensor import EmrJobFlowSensor
|
||||
|
||||
DEFAULT_ARGS = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'depends_on_past': False,
|
||||
'start_date': airflow.utils.dates.days_ago(2),
|
||||
'email': ['airflow@example.com'],
|
||||
|
|
|
@ -32,7 +32,7 @@ from airflow.contrib.operators.emr_terminate_job_flow_operator import EmrTermina
|
|||
from airflow.contrib.sensors.emr_step_sensor import EmrStepSensor
|
||||
|
||||
DEFAULT_ARGS = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'depends_on_past': False,
|
||||
'start_date': airflow.utils.dates.days_ago(2),
|
||||
'email': ['airflow@example.com'],
|
||||
|
|
|
@ -26,7 +26,7 @@ from airflow.models import DAG
|
|||
from airflow.operators.python_operator import PythonOperator
|
||||
|
||||
args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'start_date': airflow.utils.dates.days_ago(2)
|
||||
}
|
||||
|
||||
|
|
|
@ -27,7 +27,7 @@ from airflow.models import DAG
|
|||
from airflow.operators.python_operator import PythonOperator
|
||||
|
||||
default_args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'start_date': airflow.utils.dates.days_ago(2)
|
||||
}
|
||||
|
||||
|
|
|
@ -31,7 +31,7 @@ try:
|
|||
from airflow.contrib.operators.kubernetes_pod_operator import KubernetesPodOperator
|
||||
|
||||
default_args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'start_date': days_ago(2)
|
||||
}
|
||||
|
||||
|
|
|
@ -29,7 +29,7 @@ from airflow.models import DAG
|
|||
from airflow.operators.papermill_operator import PapermillOperator
|
||||
|
||||
default_args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'start_date': airflow.utils.dates.days_ago(2)
|
||||
}
|
||||
|
||||
|
|
|
@ -38,7 +38,7 @@ from airflow.operators.dummy_operator import DummyOperator
|
|||
from airflow.operators.python_operator import BranchPythonOperator, PythonOperator
|
||||
|
||||
default_args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'depends_on_past': False,
|
||||
'start_date': airflow.utils.dates.days_ago(2),
|
||||
'email': ['airflow@example.com'],
|
||||
|
|
|
@ -33,7 +33,7 @@ from airflow.contrib.sensors.qubole_sensor import QuboleFileSensor, QubolePartit
|
|||
from airflow.utils import dates
|
||||
|
||||
default_args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'depends_on_past': False,
|
||||
'start_date': dates.days_ago(2),
|
||||
'email': ['airflow@example.com'],
|
||||
|
|
|
@ -38,7 +38,7 @@ from airflow.models import DAG
|
|||
from airflow.operators.dummy_operator import DummyOperator
|
||||
|
||||
default_args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'start_date': airflow.utils.dates.days_ago(2)
|
||||
}
|
||||
|
||||
|
|
|
@ -35,7 +35,7 @@ TODO: Review the workflow, change it accordingly to
|
|||
# from airflow.operators.docker_operator import DockerOperator
|
||||
#
|
||||
# default_args = {
|
||||
# 'owner': 'Airflow',
|
||||
# 'owner': 'airflow',
|
||||
# 'depends_on_past': False,
|
||||
# 'start_date': airflow.utils.dates.days_ago(2),
|
||||
# 'email': ['airflow@example.com'],
|
||||
|
|
|
@ -27,7 +27,7 @@ from airflow.operators.bash_operator import BashOperator
|
|||
from airflow.operators.dummy_operator import DummyOperator
|
||||
|
||||
args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'start_date': airflow.utils.dates.days_ago(2),
|
||||
}
|
||||
|
||||
|
|
|
@ -27,7 +27,7 @@ from airflow.operators.dummy_operator import DummyOperator
|
|||
from airflow.operators.python_operator import BranchPythonOperator
|
||||
|
||||
args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'start_date': airflow.utils.dates.days_ago(2),
|
||||
}
|
||||
|
||||
|
|
|
@ -28,7 +28,7 @@ from airflow.operators.dummy_operator import DummyOperator
|
|||
from airflow.operators.python_operator import BranchPythonOperator
|
||||
|
||||
args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'start_date': airflow.utils.dates.days_ago(2),
|
||||
'depends_on_past': True,
|
||||
}
|
||||
|
|
|
@ -24,7 +24,7 @@ from datetime import timedelta
|
|||
from airflow.operators.docker_operator import DockerOperator
|
||||
|
||||
default_args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'depends_on_past': False,
|
||||
'start_date': airflow.utils.dates.days_ago(2),
|
||||
'email': ['airflow@example.com'],
|
||||
|
|
|
@ -25,7 +25,7 @@ from airflow.operators import bash_operator
|
|||
from airflow.operators.gcs_to_bq import GoogleCloudStorageToBigQueryOperator
|
||||
|
||||
args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'start_date': airflow.utils.dates.days_ago(2)
|
||||
}
|
||||
|
||||
|
|
|
@ -28,7 +28,7 @@ from airflow.operators.http_operator import SimpleHttpOperator
|
|||
from airflow.sensors.http_sensor import HttpSensor
|
||||
|
||||
default_args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'depends_on_past': False,
|
||||
'start_date': airflow.utils.dates.days_ago(2),
|
||||
'email': ['airflow@example.com'],
|
||||
|
|
|
@ -24,7 +24,7 @@ from airflow.models import DAG
|
|||
from airflow.operators.pig_operator import PigOperator
|
||||
|
||||
args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'start_date': airflow.utils.dates.days_ago(2),
|
||||
}
|
||||
|
||||
|
|
|
@ -27,7 +27,7 @@ from airflow.models import DAG
|
|||
from airflow.operators.python_operator import PythonOperator, PythonVirtualenvOperator
|
||||
|
||||
args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'start_date': airflow.utils.dates.days_ago(2),
|
||||
}
|
||||
|
||||
|
|
|
@ -25,7 +25,7 @@ from airflow.operators.dummy_operator import DummyOperator
|
|||
from airflow.operators.python_operator import ShortCircuitOperator
|
||||
|
||||
args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'start_date': airflow.utils.dates.days_ago(2),
|
||||
}
|
||||
|
||||
|
|
|
@ -25,7 +25,7 @@ from airflow.models import DAG
|
|||
from airflow.operators.dummy_operator import DummyOperator
|
||||
|
||||
args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'start_date': airflow.utils.dates.days_ago(2),
|
||||
}
|
||||
|
||||
|
|
|
@ -28,7 +28,7 @@ from airflow.operators.subdag_operator import SubDagOperator
|
|||
DAG_NAME = 'example_subdag_operator'
|
||||
|
||||
args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'start_date': airflow.utils.dates.days_ago(2),
|
||||
}
|
||||
|
||||
|
|
|
@ -30,7 +30,7 @@ from airflow.operators.python_operator import PythonOperator
|
|||
|
||||
dag = DAG(
|
||||
dag_id="example_trigger_target_dag",
|
||||
default_args={"start_date": airflow.utils.dates.days_ago(2), "owner": "Airflow"},
|
||||
default_args={"start_date": airflow.utils.dates.days_ago(2), "owner": "airflow"},
|
||||
schedule_interval=None,
|
||||
)
|
||||
|
||||
|
|
|
@ -24,7 +24,7 @@ from airflow import DAG
|
|||
from airflow.operators.python_operator import PythonOperator
|
||||
|
||||
args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'start_date': airflow.utils.dates.days_ago(2),
|
||||
}
|
||||
|
||||
|
|
|
@ -31,7 +31,7 @@ from airflow.operators.bash_operator import BashOperator
|
|||
# These args will get passed on to each operator
|
||||
# You can override them on a per-task basis during operator initialization
|
||||
default_args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'depends_on_past': False,
|
||||
'start_date': airflow.utils.dates.days_ago(2),
|
||||
'email': ['airflow@example.com'],
|
||||
|
|
|
@ -55,7 +55,7 @@ class DataflowCreateJavaJobOperator(BaseOperator):
|
|||
**Example**: ::
|
||||
|
||||
default_args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'depends_on_past': False,
|
||||
'start_date':
|
||||
(2016, 8, 1),
|
||||
|
|
|
@ -92,7 +92,7 @@ any of its operators. This makes it easy to apply a common parameter to many ope
|
|||
|
||||
default_args = {
|
||||
'start_date': datetime(2016, 1, 1),
|
||||
'owner': 'Airflow'
|
||||
'owner': 'airflow'
|
||||
}
|
||||
|
||||
dag = DAG('my_dag', default_args=default_args)
|
||||
|
|
|
@ -94,7 +94,7 @@ in the configuration file. When turned off, the scheduler creates a DAG run only
|
|||
|
||||
|
||||
default_args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'depends_on_past': False,
|
||||
'email': ['airflow@example.com'],
|
||||
'email_on_failure': False,
|
||||
|
|
|
@ -40,7 +40,7 @@ works.
|
|||
FILE_CATEGORIES = ["CAT1", "CAT2", "CAT3"]
|
||||
|
||||
args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'start_date': airflow.utils.dates.days_ago(2)
|
||||
}
|
||||
|
||||
|
|
|
@ -77,7 +77,7 @@ words if you have a default time zone setting of ``Europe/Amsterdam`` and create
|
|||
|
||||
default_args=dict(
|
||||
start_date=datetime(2016, 1, 1),
|
||||
owner='Airflow'
|
||||
owner='airflow'
|
||||
)
|
||||
|
||||
dag = DAG('my_dag', default_args=default_args)
|
||||
|
@ -122,7 +122,7 @@ using ``pendulum``.
|
|||
|
||||
default_args=dict(
|
||||
start_date=datetime(2016, 1, 1, tzinfo=local_tz),
|
||||
owner='Airflow'
|
||||
owner='airflow'
|
||||
)
|
||||
|
||||
dag = DAG('my_tz_dag', default_args=default_args)
|
||||
|
|
|
@ -42,7 +42,7 @@ complicated, a line by line explanation follows below.
|
|||
|
||||
|
||||
default_args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'depends_on_past': False,
|
||||
'start_date': datetime(2015, 6, 1),
|
||||
'email': ['airflow@example.com'],
|
||||
|
@ -133,7 +133,7 @@ of default parameters that we can use when creating tasks.
|
|||
from datetime import datetime, timedelta
|
||||
|
||||
default_args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'depends_on_past': False,
|
||||
'start_date': datetime(2015, 6, 1),
|
||||
'email': ['airflow@example.com'],
|
||||
|
@ -318,7 +318,7 @@ something like this:
|
|||
|
||||
|
||||
default_args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'depends_on_past': False,
|
||||
'start_date': datetime(2015, 6, 1),
|
||||
'email': ['airflow@example.com'],
|
||||
|
|
|
@ -26,7 +26,7 @@ from airflow.models import DAG
|
|||
from airflow.operators.bash_operator import BashOperator
|
||||
|
||||
args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'start_date': airflow.utils.dates.days_ago(3),
|
||||
}
|
||||
|
||||
|
|
|
@ -26,7 +26,7 @@ from airflow.models import DAG
|
|||
from airflow.operators.bash_operator import BashOperator
|
||||
|
||||
args = {
|
||||
'owner': 'Airflow',
|
||||
'owner': 'airflow',
|
||||
'start_date': airflow.utils.dates.days_ago(3),
|
||||
}
|
||||
|
||||
|
|
Загрузка…
Ссылка в новой задаче