{"id":47184474,"url":"https://github.com/a5chin/event-driven-dataflow","last_synced_at":"2026-03-13T08:36:31.045Z","repository":{"id":229418636,"uuid":"776645492","full_name":"a5chin/event-driven-dataflow","owner":"a5chin","description":"This Terraform module sets up a workflow where files stored in Cloud Storage trigger events in Eventarc, which then processes the files and stores the data in Spanner.","archived":false,"fork":false,"pushed_at":"2024-03-27T14:46:25.000Z","size":126,"stargazers_count":0,"open_issues_count":0,"forks_count":0,"subscribers_count":1,"default_branch":"main","last_synced_at":"2024-03-28T15:01:46.739Z","etag":null,"topics":["cloudfunctions","cloudstorage","dataflow","eventarc","pubsub","python3","spanner","terraform"],"latest_commit_sha":null,"homepage":"","language":"HCL","has_issues":true,"has_wiki":null,"has_pages":null,"mirror_url":null,"source_name":null,"license":null,"status":null,"scm":"git","pull_requests_enabled":true,"icon_url":"https://github.com/a5chin.png","metadata":{"files":{"readme":"README.md","changelog":null,"contributing":null,"funding":null,"license":null,"code_of_conduct":null,"threat_model":null,"audit":null,"citation":null,"codeowners":null,"security":null,"support":null,"governance":null,"roadmap":null,"authors":null,"dei":null}},"created_at":"2024-03-24T04:23:10.000Z","updated_at":"2024-03-24T14:04:20.000Z","dependencies_parsed_at":"2024-03-27T14:51:59.754Z","dependency_job_id":null,"html_url":"https://github.com/a5chin/event-driven-dataflow","commit_stats":null,"previous_names":["a5chin/event-driven-dataflow"],"tags_count":0,"template":false,"template_full_name":null,"purl":"pkg:github/a5chin/event-driven-dataflow","repository_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/a5chin%2Fevent-driven-dataflow","tags_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/a5chin%2Fevent-driven-dataflow/tags","releases_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/a5chin%2Fevent-driven-dataflow/releases","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/a5chin%2Fevent-driven-dataflow/manifests","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/a5chin","download_url":"https://codeload.github.com/a5chin/event-driven-dataflow/tar.gz/refs/heads/main","sbom_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/a5chin%2Fevent-driven-dataflow/sbom","scorecard":null,"host":{"name":"GitHub","url":"https://github.com","kind":"github","repositories_count":286080680,"owners_count":30462535,"icon_url":"https://github.com/github.png","version":null,"created_at":"2022-05-30T11:31:42.601Z","updated_at":"2026-03-13T06:34:02.089Z","status":"ssl_error","status_checked_at":"2026-03-13T06:33:49.182Z","response_time":60,"last_error":"SSL_connect returned=1 errno=0 peeraddr=140.82.121.5:443 state=error: unexpected eof while reading","robots_txt_status":"success","robots_txt_updated_at":"2025-07-24T06:49:26.215Z","robots_txt_url":"https://github.com/robots.txt","online":false,"can_crawl_api":true,"host_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories","repository_names_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repository_names","owners_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners"}},"keywords":["cloudfunctions","cloudstorage","dataflow","eventarc","pubsub","python3","spanner","terraform"],"created_at":"2026-03-13T08:36:30.546Z","updated_at":"2026-03-13T08:36:31.027Z","avatar_url":"https://github.com/a5chin.png","language":"HCL","readme":"# Terraform Module: Event-Driven batch Dataflow that imports data from Cloud Storage (GCS) to Cloud Spanner\n\n## What is this?\nThis Terraform module sets up a workflow where files stored in Cloud Storage trigger events in Eventarc, which then processes the files and stores the data in Spanner.\n\n## Architecture\n![Infrastructure Architecture](assets/images/infrastructure.png)\n1. File is stored in Cloud Storage\n2. Receive event from Eventarc\n3. If the received file is `spanner-export.json`\n4. Get Cloud API access token and\n5. POST to Dataflow creation API\n6. Read the `.avro` file in the same hierarchy as the `spanner-export.json` file and store it in Spanner\n\n## `spanner-export.json`\n```json\n{\n    \"tables\": [\n        {\n            \"name\": \"TableName\",\n            \"dataFiles\": [\n                \"TableName-000000000000.avro\",\n                \"TableName-000000000001.avro\",\n                \"TableName-[0-9]{12}.avro\"\n            ]\n        }\n    ],\n    \"dialect\":\"GOOGLE_STANDARD_SQL\"\n}\n```\n","funding_links":[],"categories":[],"sub_categories":[],"project_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fa5chin%2Fevent-driven-dataflow","html_url":"https://awesome.ecosyste.ms/projects/github.com%2Fa5chin%2Fevent-driven-dataflow","lists_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fa5chin%2Fevent-driven-dataflow/lists"}