{"id":13405680,"url":"https://github.com/scrapy/scrapy","last_synced_at":"2026-01-16T18:06:29.382Z","repository":{"id":37318576,"uuid":"529502","full_name":"scrapy/scrapy","owner":"scrapy","description":"Scrapy, a fast high-level web crawling \u0026 scraping framework for Python.","archived":false,"fork":false,"pushed_at":"2025-05-05T10:41:02.000Z","size":27849,"stargazers_count":55082,"open_issues_count":639,"forks_count":10776,"subscribers_count":1771,"default_branch":"master","last_synced_at":"2025-05-05T11:13:25.232Z","etag":null,"topics":["crawler","crawling","framework","hacktoberfest","python","scraping","web-scraping","web-scraping-python"],"latest_commit_sha":null,"homepage":"https://scrapy.org","language":"Python","has_issues":true,"has_wiki":null,"has_pages":null,"mirror_url":null,"source_name":null,"license":"bsd-3-clause","status":null,"scm":"git","pull_requests_enabled":true,"icon_url":"https://github.com/scrapy.png","metadata":{"files":{"readme":"README.rst","changelog":"NEWS","contributing":"CONTRIBUTING.md","funding":null,"license":"LICENSE","code_of_conduct":"CODE_OF_CONDUCT.md","threat_model":null,"audit":null,"citation":null,"codeowners":null,"security":"SECURITY.md","support":null,"governance":null,"roadmap":null,"authors":"AUTHORS","dei":null,"publiccode":null,"codemeta":null,"zenodo":null}},"created_at":"2010-02-22T02:01:14.000Z","updated_at":"2025-05-05T10:41:06.000Z","dependencies_parsed_at":"2024-02-05T12:49:29.240Z","dependency_job_id":"f36fe129-c231-4544-a350-2ad92989f3ff","html_url":"https://github.com/scrapy/scrapy","commit_stats":null,"previous_names":[],"tags_count":119,"template":false,"template_full_name":null,"repository_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/scrapy%2Fscrapy","tags_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/scrapy%2Fscrapy/tags","releases_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/scrapy%2Fscrapy/releases","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/scrapy%2Fscrapy/manifests","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/scrapy","download_url":"https://codeload.github.com/scrapy/scrapy/tar.gz/refs/heads/master","host":{"name":"GitHub","url":"https://github.com","kind":"github","repositories_count":252498470,"owners_count":21757810,"icon_url":"https://github.com/github.png","version":null,"created_at":"2022-05-30T11:31:42.601Z","updated_at":"2022-07-04T15:15:14.044Z","host_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories","repository_names_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repository_names","owners_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners"}},"keywords":["crawler","crawling","framework","hacktoberfest","python","scraping","web-scraping","web-scraping-python"],"created_at":"2024-07-30T19:02:08.304Z","updated_at":"2026-01-05T12:22:24.624Z","avatar_url":"https://github.com/scrapy.png","language":"Python","readme":"|logo|\n\n.. |logo| image:: https://raw.githubusercontent.com/scrapy/scrapy/master/docs/_static/logo.svg\n   :target: https://scrapy.org\n   :alt: Scrapy\n   :width: 480px\n\n|version| |python_version| |ubuntu| |macos| |windows| |coverage| |conda| |deepwiki|\n\n.. |version| image:: https://img.shields.io/pypi/v/Scrapy.svg\n   :target: https://pypi.org/pypi/Scrapy\n   :alt: PyPI Version\n\n.. |python_version| image:: https://img.shields.io/pypi/pyversions/Scrapy.svg\n   :target: https://pypi.org/pypi/Scrapy\n   :alt: Supported Python Versions\n\n.. |ubuntu| image:: https://github.com/scrapy/scrapy/workflows/Ubuntu/badge.svg\n   :target: https://github.com/scrapy/scrapy/actions?query=workflow%3AUbuntu\n   :alt: Ubuntu\n\n.. |macos| image:: https://github.com/scrapy/scrapy/workflows/macOS/badge.svg\n   :target: https://github.com/scrapy/scrapy/actions?query=workflow%3AmacOS\n   :alt: macOS\n\n.. |windows| image:: https://github.com/scrapy/scrapy/workflows/Windows/badge.svg\n   :target: https://github.com/scrapy/scrapy/actions?query=workflow%3AWindows\n   :alt: Windows\n\n.. |coverage| image:: https://img.shields.io/codecov/c/github/scrapy/scrapy/master.svg\n   :target: https://codecov.io/github/scrapy/scrapy?branch=master\n   :alt: Coverage report\n\n.. |conda| image:: https://anaconda.org/conda-forge/scrapy/badges/version.svg\n   :target: https://anaconda.org/conda-forge/scrapy\n   :alt: Conda Version\n\n.. |deepwiki| image:: https://deepwiki.com/badge.svg\n   :target: https://deepwiki.com/scrapy/scrapy\n   :alt: Ask DeepWiki\n\nScrapy_ is a web scraping framework to extract structured data from websites.\nIt is cross-platform, and requires Python 3.10+. It is maintained by Zyte_\n(formerly Scrapinghub) and `many other contributors`_.\n\n.. _many other contributors: https://github.com/scrapy/scrapy/graphs/contributors\n.. _Scrapy: https://scrapy.org/\n.. _Zyte: https://www.zyte.com/\n\nInstall with:\n\n.. code:: bash\n\n    pip install scrapy\n\nAnd follow the documentation_ to learn how to use it.\n\n.. _documentation: https://docs.scrapy.org/en/latest/\n\nIf you wish to contribute, see Contributing_.\n\n.. _Contributing: https://docs.scrapy.org/en/master/contributing.html\n","funding_links":[],"categories":["Python","Data Gathering","All","Web Crawling","Utilities","🕷️ Web Scraping Frameworks","Web 后端","Data Processing","Web爬行","Uncategorized","HarmonyOS","网络服务","爬虫","Library","framework","python","Core Libraries","Application Recommendation","web-scraping","Web Scraping \u0026 Crawling","📚 فهرست","信息获取"],"sub_categories":["Test Runners","Python","Data Pre-processing \u0026 Loading","Uncategorized","Windows Manager","网络服务_其他","5. Miscellaneous","Misc","🤖 Automation Tools","وب اسکرپینگ","爬虫与抓取框架"],"project_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fscrapy%2Fscrapy","html_url":"https://awesome.ecosyste.ms/projects/github.com%2Fscrapy%2Fscrapy","lists_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fscrapy%2Fscrapy/lists"}