This mini project shows the Air Quality Index (Industrial Air Pollution) from various locations in India in a Tableau Public dashboard.
The data is sourced from data.gov.in API using Python for cleaning & loading the data in Google Sheets.
Data is updated daily (or manually) using GitHub Actions or AWS Lambda.
My Workflow
GitHub Action "run-python.yml"
- Google credentials are stored in GitHub Actions Environment Secrets.
- data.gov.in API Key is also stored in GitHub Actions Environment Secrets.
- A GitHub Action is created that can run manually or on a schedule [12PM UTC (8PM SGT)].
- Python 3.9 is setup using actions/setup-python@v2.3.0 and the pip packages are cached.
- Dependencies are installed using py-actions/py-dependency-install@v2 based on requirements.txt (google auth, pygsheets & pandas).
- The Environment Secrets are exported to environment variables.
- Finally, the Python script is run with the environment variables passed as parameters.
Submission Category:
Wacky Wildcards
- I tried to use this workflow as a replacement / complement to the AWS Lambda function that processes the Python script at 8AM SGT.
Yaml File or Link to Code
run-python.yml
# This is a basic workflow to help you get started with Actions name: Run Python # Controls when the action will run. on: schedule: # run at 12PM UTC (8PM SGT) - cron: '0 12 * * *' # Allows you to run this workflow manually from the Actions tab workflow_dispatch: # A workflow run is made up of one or more jobs that can run sequentially or in parallel jobs: # This workflow contains a single job called "build" build: # use environment (named as "env") defined in the GitHub repository settings environment: env # The type of runner that the job will run on runs-on: ubuntu-latest # Steps represent a sequence of tasks that will be executed as part of the job steps: # Checks-out your repository under $GITHUB_WORKSPACE, so your job can access it - name: Checkout uses: actions/checkout@v2 # Set up Python 3.9 environment and cache pip packages - name: Setup Python 3.9 uses: actions/setup-python@v2.3.0 with: python-version: '3.9' cache: 'pip' # This action sets up a Python environment for use in actions by: # optionally installing and adding to PATH a version of Python that is already installed in the tools cache. # optionally caching dependencies for pip and pipenv. # Install dependencies mentioned in the requirements.txt - name: Install dependencies uses: py-actions/py-dependency-install@v2 # This GitHub Action installs Python package dependencies from a user-defined requirements.txt file path # with pip, setuptools, and wheel installs/updates during execution. # A Python package environment report is displayed at the end of Action execution. # Uses path requirements.txt and updates pip, setuptools, and wheel before the install. # Run a bash shell and store env secrets in parameters to pass to Python script - name: Get Parameters & Run Air Quality Index India Python script shell: bash env: GOVINAPIKEY: ${{ secrets.DATA_GOV_IN_API_KEY }} GDRIVEAPIKEY: ${{ secrets.GDRIVE_API_CREDENTIALS }} run: | python "Air Quality Index India.py" "$GOVINAPIKEY" "$GDRIVEAPIKEY"
Full repository is here:
AdhirKirtikar / Air-Quality-Index-India-GitHub-Actions
Repo for 2021 GitHub Actions Hackathon on DEV
Air-Quality-Index-India-GitHub-Actions
Repo for 2021 GitHub Actions Hackathon on DEV
GitHub Action "run-python.yml"
- Google credentials are stored in GitHub Actions Environment Secrets.
- data.gov.in API Key is also stored in GitHub Actions Environment Secrets.
- A GitHub Action is created that can run manually or on a schedule [12PM UTC (8PM SGT)]
- Python 3.9 is setup using actions/setup-python@v2.3.0 and the pip packages are cached
- Dependencies are installed using py-actions/py-dependency-install@v2 based on requirements.txt (google auth, pygsheets & pandas).
- The Environment Secrets are exported to environment variables.
- Finally, the Python script is run with the environment variables passed as parameters.
Python script "Air Quality Index India.py"
- The script connects to data.gov.in API using API Key passed as a parameter.
- Then it pulls the latest AQI data for India and stores in pandas dataframe.
- The data is cleaned, formatted and the columns are renamed. Nulls are replaced by 0.
- Google sheet is…
Additional Resources / Info
Tableau Public Dashboard that uses the data from the generated Google Sheets: "Air Quality - Pollutant Index - India"
Top comments (0)