{"id":26922586,"url":"https://github.com/microsoft/DeepSpeed","last_synced_at":"2025-04-02T00:04:07.172Z","repository":{"id":37055596,"uuid":"235860204","full_name":"deepspeedai/DeepSpeed","owner":"deepspeedai","description":"DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.","archived":false,"fork":false,"pushed_at":"2025-03-31T23:00:03.000Z","size":225069,"stargazers_count":37696,"open_issues_count":1133,"forks_count":4324,"subscribers_count":348,"default_branch":"master","last_synced_at":"2025-03-31T23:04:23.914Z","etag":null,"topics":["billion-parameters","compression","data-parallelism","deep-learning","gpu","inference","machine-learning","mixture-of-experts","model-parallelism","pipeline-parallelism","pytorch","trillion-parameters","zero"],"latest_commit_sha":null,"homepage":"https://www.deepspeed.ai/","language":"Python","has_issues":true,"has_wiki":null,"has_pages":null,"mirror_url":null,"source_name":null,"license":"apache-2.0","status":null,"scm":"git","pull_requests_enabled":true,"icon_url":"https://github.com/deepspeedai.png","metadata":{"files":{"readme":"README.md","changelog":null,"contributing":"CONTRIBUTING.md","funding":null,"license":"LICENSE","code_of_conduct":"CODE_OF_CONDUCT.md","threat_model":null,"audit":null,"citation":null,"codeowners":"CODEOWNERS","security":"SECURITY.md","support":null,"governance":"GOVERNANCE.md","roadmap":null,"authors":null,"dei":null,"publiccode":null,"codemeta":null}},"created_at":"2020-01-23T18:35:18.000Z","updated_at":"2025-03-31T21:25:32.000Z","dependencies_parsed_at":"2023-09-28T17:40:23.869Z","dependency_job_id":"7f1ee5f6-71f0-45e8-a40d-89c867961628","html_url":"https://github.com/deepspeedai/DeepSpeed","commit_stats":{"total_commits":2523,"total_committers":370,"mean_commits":6.818918918918919,"dds":0.8252080856123662,"last_synced_commit":"6eefc3d0ead2c6360eec248daab0cae66a737ea0"},"previous_names":["deepspeedai/deepspeed","microsoft/deepspeed"],"tags_count":99,"template":false,"template_full_name":null,"repository_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/deepspeedai%2FDeepSpeed","tags_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/deepspeedai%2FDeepSpeed/tags","releases_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/deepspeedai%2FDeepSpeed/releases","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/deepspeedai%2FDeepSpeed/manifests","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/deepspeedai","download_url":"https://codeload.github.com/deepspeedai/DeepSpeed/tar.gz/refs/heads/master","host":{"name":"GitHub","url":"https://github.com","kind":"github","repositories_count":246730309,"owners_count":20824399,"icon_url":"https://github.com/github.png","version":null,"created_at":"2022-05-30T11:31:42.601Z","updated_at":"2022-07-04T15:15:14.044Z","host_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories","repository_names_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repository_names","owners_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners"}},"keywords":["billion-parameters","compression","data-parallelism","deep-learning","gpu","inference","machine-learning","mixture-of-experts","model-parallelism","pipeline-parallelism","pytorch","trillion-parameters","zero"],"created_at":"2025-04-02T00:03:16.408Z","updated_at":"2025-04-02T00:04:02.145Z","avatar_url":"https://github.com/deepspeedai.png","language":"Python","readme":null,"funding_links":[],"categories":["Python","🌟 Awesome Lists and Resource Hubs","Software","Optimizations and fine-tuning","Frameworks for Training","Model Training \u0026 Experiments","Fine-Tuning \u0026 Training","Collections","Frameworks","Transformers and LLMs","Frameworks-for-Training","LLM Training Frameworks","Projects","pytorch","精选开源项目合集","MLVU Frameworks","分布式机器学习","Python / C++","Training","🌗 Model Scalability","Deep Learning Framework","LLM训练框架","NLP","Compression System:","Repos","Tools","Model Serving Frameworks","Application","MoE Application","Libraries","Tools (GitHub)","Projects and Tools","Optimization Tools","🛠️ Libraries","📋 Contents","Deep Learning","1. **Large Language Models**","Fine-tuning \u0026 Quantization (18)","📋 List of Open-Source Projects","9. Fine-Tuning","Training \u0026 Fine-Tuning","LLMs Training","Training and Fine-Tuning"],"sub_categories":["Trends","Training Infrastructure","Useful Repositories","Survey","Frameworks and Libraries","Popular-LLM","GPT开源平替机器人🔥🔥🔥","Weakly-supervised learning","Mojo🔥FastAPI Client","GPT开源平替机器人","Frameworks for Training","Deployment \u0026 Distribution","LLM 评估与数据","3. Pretraining","Runtimes/Deployment","LangManus","2018","Inference Framework","LLMOps vs MLOps","LLM (Large Language Model)","📚 arXiv","🧬 1. Core Frameworks \u0026 Libraries","AI Infrastructure","LLM Infra and Optimization","Training Frameworks","Libraries"],"project_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fmicrosoft%2FDeepSpeed","html_url":"https://awesome.ecosyste.ms/projects/github.com%2Fmicrosoft%2FDeepSpeed","lists_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fmicrosoft%2FDeepSpeed/lists"}