Data engineering to analyze Earthquake events - ETL project using DBT, Azure ADF, Snowflake and Tableau
USGS.GOV - https://earthquake.usgs.gov/fdsnws/event/1 The website provides daily activities of earthquake events recorded
Using the Azure data factory copy activity, the data is scheduled to load every day to Snowflake CSV data is extracted using REST API and loaded to the Snowflake warehouse

Used as a data warehouse to store the raw data and connected to DBT for transformations the updated data is merged into the final table, Tableau uses the final table to update the dashboard.
DBT cloud is used for data transformations and data cleaning, countries data is extracted from lat and log values and added after applying the transformations the changes are merged into Snowflake. DBT is also scheduled to run every day.
Tableau dashboard with data visualizations and graphs - https://public.tableau.com/app/profile/solomon8607/viz/EarthquakeEvents/Dashboard1
Screenshots:
