{"id":13409319,"url":"https://github.com/NVIDIA/TensorRT-LLM","last_synced_at":"2025-03-14T14:31:17.923Z","repository":{"id":201227926,"uuid":"679366051","full_name":"NVIDIA/TensorRT-LLM","owner":"NVIDIA","description":"TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.","archived":false,"fork":false,"pushed_at":"2025-03-14T02:23:55.000Z","size":952171,"stargazers_count":9701,"open_issues_count":518,"forks_count":1149,"subscribers_count":102,"default_branch":"main","last_synced_at":"2025-03-14T03:35:25.502Z","etag":null,"topics":[],"latest_commit_sha":null,"homepage":"https://nvidia.github.io/TensorRT-LLM","language":"C++","has_issues":true,"has_wiki":null,"has_pages":null,"mirror_url":null,"source_name":null,"license":"apache-2.0","status":null,"scm":"git","pull_requests_enabled":true,"icon_url":"https://github.com/NVIDIA.png","metadata":{"files":{"readme":"README.md","changelog":null,"contributing":null,"funding":null,"license":"LICENSE","code_of_conduct":null,"threat_model":null,"audit":null,"citation":null,"codeowners":null,"security":null,"support":null,"governance":null,"roadmap":null,"authors":null,"dei":null,"publiccode":null,"codemeta":null}},"created_at":"2023-08-16T17:14:27.000Z","updated_at":"2025-03-14T00:10:48.000Z","dependencies_parsed_at":"2025-02-28T03:41:13.573Z","dependency_job_id":null,"html_url":"https://github.com/NVIDIA/TensorRT-LLM","commit_stats":{"total_commits":120,"total_committers":19,"mean_commits":6.315789473684211,"dds":0.3416666666666667,"last_synced_commit":"be1788106245496872d18e702978e59b6bfd50e0"},"previous_names":["nvidia/tensorrt-llm"],"tags_count":17,"template":false,"template_full_name":null,"repository_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/NVIDIA%2FTensorRT-LLM","tags_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/NVIDIA%2FTensorRT-LLM/tags","releases_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/NVIDIA%2FTensorRT-LLM/releases","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/NVIDIA%2FTensorRT-LLM/manifests","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/NVIDIA","download_url":"https://codeload.github.com/NVIDIA/TensorRT-LLM/tar.gz/refs/heads/main","host":{"name":"GitHub","url":"https://github.com","kind":"github","repositories_count":243593358,"owners_count":20316172,"icon_url":"https://github.com/github.png","version":null,"created_at":"2022-05-30T11:31:42.601Z","updated_at":"2022-07-04T15:15:14.044Z","host_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories","repository_names_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repository_names","owners_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners"}},"keywords":[],"created_at":"2024-07-30T20:00:59.851Z","updated_at":"2025-03-14T14:31:12.895Z","avatar_url":"https://github.com/NVIDIA.png","language":"C++","readme":null,"funding_links":[],"categories":["Models and Tools","Serving","A01_文本生成_文本对话","C++","LLM Deployment","Tools","Inference \u0026 Deployment","\u003ca name=\"cpp\"\u003e\u003c/a\u003eC++","Python","推理 Inference","LLM inference engines","Summary","Lighter and Deployment Frameworks","Awesome Open-Sourced LLMSys Projects","Inference","开发者工具 \u0026 AI Infra","NLP","Inference engines","🔓 Open Source Inference Engines","LLM Inference","🚀 **Inference Frameworks \u0026 Runtimes**","Repos","⚙️ Systems and Multi-GPU Engineering","Frameworks","Model Serving Frameworks","Inference Engine","Open-Source Local LLM Projects","🛠️ AI 工具与框架","🖥 Local Deployment Tools","📋 Contents","Deployment","2. **Production Tools**","LLM部署与本地运行","Inference Engines \u0026 Backends (22)","LLM Serving / Inference","8. Inference Engines","Inference \u0026 Serving","📱 Edge \u0026 On-Device AI","Local Inference and Serving"],"sub_categories":["LLM Deployment","Large Model Serving","大语言对话模型及数据","Other","Cloud \u0026 Container Deployment","Distributed Systems","推理与部署","3. Pretraining","⚡ **TensorRT-LLM**","LangManus","LLM 推理与部署","Server Deployment \u0026 High-Performance Inference","⚡ 3. Inference Engines \u0026 Serving","LLM 评估与数据","Server / Production","Inference Engines","Notable MCP Servers","Serve at scale"],"project_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2FNVIDIA%2FTensorRT-LLM","html_url":"https://awesome.ecosyste.ms/projects/github.com%2FNVIDIA%2FTensorRT-LLM","lists_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2FNVIDIA%2FTensorRT-LLM/lists"}