Attribute |
Value |
dag_id |
crawlers_hourly |
duration |
None |
end_date |
2025-04-20 20:00:02.758987+00:00 |
execution_date |
2025-04-20T19:00:00+00:00 |
executor_config |
{} |
generate_command |
<function TaskInstance.generate_command at 0x7f2ad4931b70> |
hostname |
68bb9f6dae8e |
is_premature |
False |
job_id |
None |
key |
('crawlers_hourly', 'Start', <Pendulum [2025-04-20T19:00:00+00:00]>, 2) |
log |
<Logger airflow.task (INFO)> |
log_filepath |
/usr/local/airflow/logs/crawlers_hourly/Start/2025-04-20T19:00:00+00:00.log |
log_url |
http://localhost:8080/admin/airflow/log?dag_id=crawlers_hourly&task_id=Start&execution_date=2025-04-20T19%3A00%3A00%2B00%3A00 |
logger |
<Logger airflow.task (INFO)> |
mark_success_url |
http://localhost:8080/admin/airflow/success?task_id=Start&dag_id=crawlers_hourly&execution_date=2025-04-20T19%3A00%3A00%2B00%3A00&upstream=false&downstream=false |
max_tries |
2 |
metadata |
MetaData(bind=None) |
next_try_number |
2 |
operator |
None |
pid |
705903 |
pool |
general |
previous_ti |
<TaskInstance: crawlers_hourly.Start 2025-04-20 18:00:00+00:00 [success]> |
priority_weight |
7 |
queue |
default |
queued_dttm |
None |
raw |
False |
run_as_user |
None |
start_date |
2025-04-20 20:00:01.809806+00:00 |
state |
success |
task |
<Task(DummyOperator): Start> |
task_id |
Start |
test_mode |
False |
try_number |
2 |
unixname |
airflow |
Attribute |
Value |
adhoc |
False |
dag |
<DAG: crawlers_hourly> |
dag_id |
crawlers_hourly |
depends_on_past |
True |
deps |
{<TIDep(Trigger Rule)>, <TIDep(Not In Retry Period)>, <TIDep(Previous Dagrun State)>} |
downstream_list |
[<Task(DockerOperator): linkup-crawl>, <Task(DockerOperator): bcjobs-crawl>, <Task(DockerOperator): linkedin-crawl>, <Task(DockerOperator): simplyhired-crawl>] |
downstream_task_ids |
{'linkup-crawl', 'bcjobs-crawl', 'linkedin-crawl', 'simplyhired-crawl'} |
email |
['airflow@airflow.com'] |
email_on_failure |
False |
email_on_retry |
False |
end_date |
None |
execution_timeout |
None |
executor_config |
{} |
inlets |
[] |
lineage_data |
None |
log |
<Logger airflow.task.operators (INFO)> |
logger |
<Logger airflow.task.operators (INFO)> |
max_retry_delay |
None |
on_failure_callback |
None |
on_retry_callback |
None |
on_success_callback |
None |
outlets |
[] |
owner |
airflow |
params |
{} |
pool |
general |
priority_weight |
1 |
priority_weight_total |
7 |
queue |
default |
resources |
{'cpus': {'_name': 'CPU', '_units_str': 'core(s)', '_qty': 1}, 'ram': {'_name': 'RAM', '_units_str': 'MB', '_qty': 512}, 'disk': {'_name': 'Disk', '_units_str': 'MB', '_qty': 512}, 'gpus': {'_name': 'GPU', '_units_str': 'gpu(s)', '_qty': 0}} |
retries |
2 |
retry_delay |
0:30:00 |
retry_exponential_backoff |
False |
run_as_user |
None |
schedule_interval |
0 * * * * |
sla |
None |
start_date |
2020-12-17T00:00:00+00:00 |
task_concurrency |
None |
task_id |
Start |
task_type |
DummyOperator |
template_ext |
[] |
template_fields |
() |
trigger_rule |
all_success |
ui_color |
#e8f7e4 |
ui_fgcolor |
#000 |
upstream_list |
[] |
upstream_task_ids |
set() |
wait_for_downstream |
False |
weight_rule |
downstream |