{"id":17860955,"url":"https://github.com/mazzasaverio/doccrawl","last_synced_at":"2026-04-11T07:43:15.267Z","repository":{"id":259705862,"uuid":"879233666","full_name":"mazzasaverio/doccrawl","owner":"mazzasaverio","description":"Simple document crawler that harvests PDFs and documents from configured web sources.","archived":false,"fork":false,"pushed_at":"2024-11-12T10:00:49.000Z","size":514,"stargazers_count":1,"open_issues_count":0,"forks_count":0,"subscribers_count":1,"default_branch":"master","last_synced_at":"2025-08-14T13:42:11.323Z","etag":null,"topics":["asyncpg","data-engineering","docker","logfire","playwright","postgresql","pydantic-v2","python3","scrapegraphai"],"latest_commit_sha":null,"homepage":"","language":"Python","has_issues":true,"has_wiki":null,"has_pages":null,"mirror_url":null,"source_name":null,"license":"mit","status":null,"scm":"git","pull_requests_enabled":true,"icon_url":"https://github.com/mazzasaverio.png","metadata":{"files":{"readme":"README.md","changelog":null,"contributing":null,"funding":null,"license":"LICENSE","code_of_conduct":null,"threat_model":null,"audit":null,"citation":null,"codeowners":null,"security":null,"support":null,"governance":null,"roadmap":null,"authors":null,"dei":null,"publiccode":null,"codemeta":null}},"created_at":"2024-10-27T11:23:24.000Z","updated_at":"2024-11-12T10:00:53.000Z","dependencies_parsed_at":"2024-10-27T13:41:04.090Z","dependency_job_id":"c51a5704-426e-435a-aeca-101b12b883e9","html_url":"https://github.com/mazzasaverio/doccrawl","commit_stats":{"total_commits":10,"total_committers":2,"mean_commits":5.0,"dds":0.09999999999999998,"last_synced_commit":"3107ccb0910e3f0137c20efefce47fe405d864cd"},"previous_names":["mazzasaverio/doccrawl"],"tags_count":0,"template":false,"template_full_name":null,"purl":"pkg:github/mazzasaverio/doccrawl","repository_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/mazzasaverio%2Fdoccrawl","tags_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/mazzasaverio%2Fdoccrawl/tags","releases_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/mazzasaverio%2Fdoccrawl/releases","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/mazzasaverio%2Fdoccrawl/manifests","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/mazzasaverio","download_url":"https://codeload.github.com/mazzasaverio/doccrawl/tar.gz/refs/heads/master","sbom_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/mazzasaverio%2Fdoccrawl/sbom","scorecard":null,"host":{"name":"GitHub","url":"https://github.com","kind":"github","repositories_count":286080680,"owners_count":31673067,"icon_url":"https://github.com/github.png","version":null,"created_at":"2022-05-30T11:31:42.601Z","updated_at":"2026-04-10T17:19:37.612Z","status":"online","status_checked_at":"2026-04-11T02:00:05.776Z","response_time":54,"last_error":null,"robots_txt_status":"success","robots_txt_updated_at":"2025-07-24T06:49:26.215Z","robots_txt_url":"https://github.com/robots.txt","online":true,"can_crawl_api":true,"host_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories","repository_names_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repository_names","owners_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners"}},"keywords":["asyncpg","data-engineering","docker","logfire","playwright","postgresql","pydantic-v2","python3","scrapegraphai"],"created_at":"2024-10-28T08:03:45.579Z","updated_at":"2026-04-11T07:43:15.223Z","avatar_url":"https://github.com/mazzasaverio.png","language":"Python","readme":"# Scrapy Frontier Crawler\n\nA configurable web crawler built with Scrapy and Playwright for handling both static and dynamic content. The crawler can process different types of URLs and store results in a PostgreSQL database.\n\n## Features\n\n- 🔍 Three types of URL processing:\n  - Type 0: Direct target URL processing\n  - Type 1: Static page scanning for target URLs\n  - Type 2: Dynamic page scanning with depth navigation\n- 🎭 Playwright integration for JavaScript-rendered content\n- 📊 PostgreSQL storage for crawled URLs and stats \n- 🔧 YAML-based configuration\n- 📝 Structured logging with Logfire\n- 🐳 Docker support\n- ☁️ Azure deployment ready with Terraform\n\n## Prerequisites\n\n- Python 3.11+\n- PostgreSQL database\n- [uv](https://github.com/astral-sh/uv) for package management\n- Docker (optional)\n\n\n## Contributing\n\n1. Fork the repository\n2. Create a feature branch\n3. Commit your changes\n4. Push to the branch\n5. Create a Pull Request\n\n","funding_links":[],"categories":[],"sub_categories":[],"project_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fmazzasaverio%2Fdoccrawl","html_url":"https://awesome.ecosyste.ms/projects/github.com%2Fmazzasaverio%2Fdoccrawl","lists_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fmazzasaverio%2Fdoccrawl/lists"}