You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This repo contains two pipelines for collecting and processing podcast metrics data
from Spotify, Apple, Podigee, and Anchor. It uses the following connector libs:
A pipeline consists of a Docker image which uses cron to run the connector periodically
and forwards the fetched data to the Open Podcast API.
Getting started
Starting the full stack
For local development, you can use the docker-compose.yml file to run the pipeline.
To start the stack, consisting of the pipeline and a sample database, run:
make up
This will start the pipeline and a sample database. The pipeline will run every
minute and fetch data from Spotify, Apple, and Anchor according to the
podcastSources SQL table (see db_local_dev/schema.sql).
Starting the database only
Alternatively, you can also start the database only.
This is helpful if you want to run the pipeline in interactive mode
docker-compose up db
After that, you can start the manager container to run the pipeline:
# Interactive shell
docker-compose run --rm --entrypoint /bin/bash --env-file connector_manager/.env manager
# Run the pipeline in interactive mode
python -m manager --interactive