Welcome to our AI startup's ambitious project aimed at transforming the way traffic data is analyzed and utilized for smart city initiatives. This README will guide you through the project structure, business need, data source, and instructions for setting up the data warehouse tech stack using Airflow, dbt, PostgreSQL, and Redash.
- /dags: Airflow DAG scripts for orchestrating data loading and transformation.
- /dwh_dbt: dbt project folder containing models for data transformation.
- /notebooks: Jupyter notebook files for raw CSV data processing and loading.
- /data: Raw CSV files used in the project.
- /images: Images used in the README.
- /redash: Redash project folder for visualization and reporting.
- /test: Contains test scripts.
- .gitignore: Git configuration to exclude unnecessary files.
- docker-compose.yaml: Docker Compose configuration for fully dockerized deployment.
- requirements.txt: File containing the list of project dependencies.
Our AI startup collaborates with businesses to deploy sensors, collecting diverse data for critical intelligence. The city traffic department has entrusted us to create a scalable data warehouse that analyzes vehicle trajectory data from swarm UAVs, aiming to enhance traffic flow and contribute to undisclosed projects.
We utilize the pNEUMA dataset, a large-scale collection of naturalistic vehicle trajectories in Athens, Greece. This dataset, acquired by a unique experiment using swarm drones, provides valuable insights into traffic patterns.
Data Source: pNEUMA Dataset
References for Understanding Data Generation:
- PIA15_poster.pdf
- Automatic vehicle trajectory extraction for traffic analysis from aerial video data
Visualization and Interaction Tools:
- travia: Traffic data Visualization and Annotation tool
- pNEUMA_mastersproject: Python files to work with pNEUMA dataset
-
Clone the Repository:
git clone https://github.com/your-username/Data-Warehouse.git cd Data-Warehouse
-
Install Dependencies:
pip install -r requirements.txt
-
Run Airflow Services:
docker-compose up --build
-
Access Airflow UI:
- Open your browser and go to http://localhost:8080.
-
Customize DAGs and dbt Models:
- Adjust Airflow DAGs in the dags folder.
- Modify dbt models in the dwh_dbt folder based on your specific requirements and data transformations.
-
Stop Airflow Services:
docker-compose down
This project aligns with our mission to revolutionize traffic analytics, providing cities with actionable insights for smart urban planning. We invite you to explore, contribute, and leverage our data warehouse solution to make impactful decisions for a more connected and efficient future.