Create a pipeline to fetch weather data from an API, process it, and store it in a database for further analysis and visualization.
You can check my presentation with this LINK
- Programming Language: Python (for scripting and data manipulation)
- API: OpenWeatherMap API
- Database: PostgreSQL
- Tools: Docker (for containerization), Airflow (for scheduling and orchestrating the pipeline), Grafana (for visualization)
- Dockerize your environment to ensure consistency and portability.
- Set up containers for PostgreSQL (for data storage), Airflow (for orchestration), and Grafana (for visualization).
- {make sure the db is auto-build and the coordinates data is already in db}
- {how to append collected data in development}
- You can Sign Up on Weather API to get your API keys
- {How to set user-own API}
- {installing airflow}
- {make sure the DAGs is already have}
- {make sure connected to postgres}
- {How to grafana}
- {make sure db is configurated}
- {ensure dashboard template is exist}
- Security: Ensure API keys and sensitive data are securely managed within your Docker/local environment.
By building this pipeline, I gained practical experience in fetching API, scheduling DAGs, Processing raw data, also evaluate and monitoring data by visualization in dashboard
- Make improvement about Data Governance in the database
- Adding more API or use real hardware so it can be streamed with real-time data and much more source data
(16/07/24)
- Fixing README.md
TODO :
- Make sure docker and build makefile to initialize container
- Explaining how the code works
- Make sure the image, database, and API key to smoothly runs