{"id":19542039,"url":"https://github.com/bigscience-workshop/interpretability-ideas","last_synced_at":"2026-03-19T10:22:01.145Z","repository":{"id":103065742,"uuid":"461896439","full_name":"bigscience-workshop/interpretability-ideas","owner":"bigscience-workshop","description":null,"archived":false,"fork":false,"pushed_at":"2022-02-27T08:57:07.000Z","size":19,"stargazers_count":1,"open_issues_count":10,"forks_count":1,"subscribers_count":24,"default_branch":"master","last_synced_at":"2025-09-09T16:09:34.408Z","etag":null,"topics":[],"latest_commit_sha":null,"homepage":null,"language":null,"has_issues":true,"has_wiki":null,"has_pages":null,"mirror_url":null,"source_name":null,"license":null,"status":null,"scm":"git","pull_requests_enabled":true,"icon_url":"https://github.com/bigscience-workshop.png","metadata":{"files":{"readme":"README.md","changelog":null,"contributing":null,"funding":null,"license":null,"code_of_conduct":null,"threat_model":null,"audit":null,"citation":null,"codeowners":null,"security":null,"support":null,"governance":null,"roadmap":null,"authors":null,"dei":null,"publiccode":null,"codemeta":null}},"created_at":"2022-02-21T14:27:16.000Z","updated_at":"2023-03-22T00:47:38.000Z","dependencies_parsed_at":null,"dependency_job_id":"f706020f-3dc6-45e2-b61c-62cd43f5dde7","html_url":"https://github.com/bigscience-workshop/interpretability-ideas","commit_stats":null,"previous_names":[],"tags_count":0,"template":false,"template_full_name":null,"purl":"pkg:github/bigscience-workshop/interpretability-ideas","repository_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/bigscience-workshop%2Finterpretability-ideas","tags_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/bigscience-workshop%2Finterpretability-ideas/tags","releases_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/bigscience-workshop%2Finterpretability-ideas/releases","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/bigscience-workshop%2Finterpretability-ideas/manifests","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/bigscience-workshop","download_url":"https://codeload.github.com/bigscience-workshop/interpretability-ideas/tar.gz/refs/heads/master","sbom_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/bigscience-workshop%2Finterpretability-ideas/sbom","scorecard":null,"host":{"name":"GitHub","url":"https://github.com","kind":"github","repositories_count":286080680,"owners_count":29986239,"icon_url":"https://github.com/github.png","version":null,"created_at":"2022-05-30T11:31:42.601Z","updated_at":"2026-03-01T21:06:37.093Z","status":"ssl_error","status_checked_at":"2026-03-01T21:05:45.052Z","response_time":124,"last_error":"SSL_read: unexpected eof while reading","robots_txt_status":"success","robots_txt_updated_at":"2025-07-24T06:49:26.215Z","robots_txt_url":"https://github.com/robots.txt","online":false,"can_crawl_api":true,"host_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories","repository_names_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repository_names","owners_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners"}},"keywords":[],"created_at":"2024-11-11T03:12:54.194Z","updated_at":"2026-03-01T22:01:28.384Z","avatar_url":"https://github.com/bigscience-workshop.png","language":null,"readme":"# BigScience @ GSoC 2022\n\u003cimg align=\"right\" height=\"27%\" width=\"27%\" src=\"https://images.squarespace-cdn.com/content/v1/616206d72ac74d0c65656167/4aadf3be-fea0-489d-bfee-4f376e724eb7/sakura-flower-clipart-md-spaces.png?format=250w\"/\u003e\n\n**Welcome to our BigScience Interpretability Ideas repo for the GSoC 2022!**\n\n\nDuring the BigScience workshop, we've established a community of collaborators from all around the world. \nA specific group of the BigScience community developed a [vision](https://youtu.be/NL1_kMOkHm8?t=647) for the exact steps needed to improve the interpretability of NLP models. \nThe implementation of these interpretability insights are the core idea of our Summer of Code! \nOur ideas are backed by the one-year-long research and are in line with the ongoing projects.\n\nBigScience workshop is community-driven by design, and as the newcomer contributors you will [join the community](https://bigscience.notion.site/Information-How-to-join-2519d6a2c6054af5976c76999ebf72e3) once you start working on proposals.\nDuring the GSoC, we expect contributors to follow the community groups' agenda and attend the sync calls together with all the group contributors.\n\n**Reach out to us on the [bigscience-gsoc@googlegroups.com](bigscience-gsoc@googlegroups.com)** with general questions!\u003c/br\u003e\n**Discuss and propose the [Ideas](https://github.com/bigscience-workshop/interpretability-gsoc-ideas/issues)** in the [Issues](https://github.com/bigscience-workshop/interpretability-gsoc-ideas/issues) threads here!\n\n---\n\n\u003cp align=\"right\"\u003e\u003cb\u003eRead more about \u003ca href=\"https://bigscience.huggingface.co\"\u003eBigScience\u003c/a\u003e\u003c/b\u003e\u003c/p\u003e\n","funding_links":[],"categories":[],"sub_categories":[],"project_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fbigscience-workshop%2Finterpretability-ideas","html_url":"https://awesome.ecosyste.ms/projects/github.com%2Fbigscience-workshop%2Finterpretability-ideas","lists_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fbigscience-workshop%2Finterpretability-ideas/lists"}