{"id":22058512,"url":"https://github.com/Infini-AI-Lab/MagicPIG","last_synced_at":"2025-07-24T00:31:43.107Z","repository":{"id":259413685,"uuid":"876363500","full_name":"Infini-AI-Lab/MagicPIG","owner":"Infini-AI-Lab","description":"MagicPIG: LSH Sampling for Efficient LLM Generation","archived":false,"fork":false,"pushed_at":"2024-10-24T09:31:12.000Z","size":56962,"stargazers_count":27,"open_issues_count":0,"forks_count":0,"subscribers_count":1,"default_branch":"main","last_synced_at":"2024-10-25T07:09:31.741Z","etag":null,"topics":["decoding","gpu-cpu","llm-inference","lsh-algorithm"],"latest_commit_sha":null,"homepage":"","language":"Python","has_issues":true,"has_wiki":null,"has_pages":null,"mirror_url":null,"source_name":null,"license":"apache-2.0","status":null,"scm":"git","pull_requests_enabled":true,"icon_url":"https://github.com/Infini-AI-Lab.png","metadata":{"files":{"readme":"README.md","changelog":null,"contributing":null,"funding":null,"license":"LICENSE","code_of_conduct":null,"threat_model":null,"audit":null,"citation":null,"codeowners":null,"security":null,"support":null,"governance":null,"roadmap":null,"authors":null,"dei":null,"publiccode":null,"codemeta":null}},"created_at":"2024-10-21T20:57:26.000Z","updated_at":"2024-10-25T05:33:30.000Z","dependencies_parsed_at":"2024-10-25T11:06:27.845Z","dependency_job_id":null,"html_url":"https://github.com/Infini-AI-Lab/MagicPIG","commit_stats":null,"previous_names":["infini-ai-lab/magicpig"],"tags_count":0,"template":false,"template_full_name":null,"purl":"pkg:github/Infini-AI-Lab/MagicPIG","repository_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/Infini-AI-Lab%2FMagicPIG","tags_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/Infini-AI-Lab%2FMagicPIG/tags","releases_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/Infini-AI-Lab%2FMagicPIG/releases","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/Infini-AI-Lab%2FMagicPIG/manifests","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/Infini-AI-Lab","download_url":"https://codeload.github.com/Infini-AI-Lab/MagicPIG/tar.gz/refs/heads/main","sbom_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/Infini-AI-Lab%2FMagicPIG/sbom","host":{"name":"GitHub","url":"https://github.com","kind":"github","repositories_count":266774717,"owners_count":23982246,"icon_url":"https://github.com/github.png","version":null,"created_at":"2022-05-30T11:31:42.601Z","updated_at":"2022-07-04T15:15:14.044Z","status":"online","status_checked_at":"2025-07-23T02:00:09.312Z","response_time":66,"last_error":null,"robots_txt_status":null,"robots_txt_updated_at":null,"robots_txt_url":"https://github.com/robots.txt","online":true,"can_crawl_api":true,"host_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories","repository_names_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repository_names","owners_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners"}},"keywords":["decoding","gpu-cpu","llm-inference","lsh-algorithm"],"created_at":"2024-11-30T17:02:12.405Z","updated_at":"2025-07-24T00:31:38.088Z","avatar_url":"https://github.com/Infini-AI-Lab.png","language":"Python","readme":null,"funding_links":[],"categories":["A01_文本生成_文本对话","Python"],"sub_categories":["大语言对话模型及数据"],"project_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2FInfini-AI-Lab%2FMagicPIG","html_url":"https://awesome.ecosyste.ms/projects/github.com%2FInfini-AI-Lab%2FMagicPIG","lists_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2FInfini-AI-Lab%2FMagicPIG/lists"}