Skip to content
Snippets Groups Projects
Commit 60acd559 authored by Afnan  Ramadhan's avatar Afnan Ramadhan
Browse files

Merge branch 'fix/create-connection-in-airflow' into 'develop'

fix: add connection airflow

See merge request !6
parents 59f758f9 a7969dcd
Branches
2 merge requests!10chore: add model version control and finally can trigger the sparking flow...,!6fix: add connection airflow
Pipeline #66411 failed with stages
in 19 seconds
stages:
- build
- run
- clean-up
build:
stage: build
......@@ -23,3 +24,16 @@ run:
script:
- echo "Run Docker Container"
- docker compose up -d
clean-up:
stage: clean-up
tags:
- macos
only:
- main
- develop
script:
- echo "Clean up "
- CONTAINER_ID=$(docker ps -q --filter "name=webserver-1")
- docker exec -it $CONTAINER_ID airflow dags trigger train_model
......@@ -31,7 +31,7 @@ def check_accuracy(**kwargs):
accuracy_data = Variable.get("data_drift_accuracy", deserialize_json=True)
accuracy = accuracy_data.get("accuracy", 0)
if accuracy < 0.70:
return 'trigger_sparking_flow'
return 'trigger_train_model'
else:
return 'end'
......@@ -42,9 +42,9 @@ branch_task = PythonOperator(
dag=dag,
)
trigger_sparking_flow = TriggerDagRunOperator(
task_id="trigger_sparking_flow",
trigger_dag_id="sparking_flow",
trigger_train_model = TriggerDagRunOperator(
task_id="trigger_train_model",
trigger_dag_id="train_model",
dag=dag,
)
......@@ -55,5 +55,5 @@ end = PythonOperator(
)
start >> data_drift_job >> branch_task
branch_task >> trigger_sparking_flow >> end
branch_task >> trigger_train_model >> end
branch_task >> end
......@@ -4,7 +4,7 @@ from airflow.operators.python import PythonOperator
from airflow.providers.apache.spark.operators.spark_submit import SparkSubmitOperator
dag = DAG(
dag_id = "sparking_flow",
dag_id = "train_model",
default_args = {
"owner": "admin",
"start_date": airflow.utils.dates.days_ago(1)
......
......@@ -40,11 +40,11 @@ services:
depends_on:
- spark-master
environment:
SPARK_MODE: worker
SPARK_WORKER_CORES: 2
SPARK_WORKER_MEMORY: 1g
SPARK_MASTER_URL: spark://spark-master:7077
SPARK_MASTER: spark://spark-master:7077
- SPARK_MODE=worker
- SPARK_WORKER_CORES=2
- SPARK_WORKER_MEMORY=1g
- SPARK_MASTER_URL=spark://spark-master:7077
- SPARK_MASTER=spark://spark-master:7077
postgres:
image: postgres:14.0
......@@ -57,7 +57,15 @@ services:
webserver:
<<: *airflow-common
command: bash -c "airflow db init && airflow webserver"
command: >
bash -c "
airflow db init &&
if airflow connections get spark_conn; then
airflow connections delete spark_conn;
fi;
airflow connections add 'spark_conn' --conn-type 'spark' --conn-host 'spark://spark-master' --conn-port '7077' --conn-extra '{\"queue\": \"\", \"deploy_mode\": \"client\", \"spark_binary\": \"spark-submit\"}';
rm -f /opt/airflow/airflow-webserver.pid;
airflow webserver"
ports:
- "8080:8080"
depends_on:
......@@ -68,4 +76,5 @@ services:
command: bash -c "airflow db init && airflow db migrate && airflow users create --username admin --firstname Admin --lastname Admin --role Admin --email admin@example.com --password admin && airflow scheduler"
networks:
airflow:
\ No newline at end of file
airflow:
driver: bridge
\ No newline at end of file
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment