{"id":27630331,"url":"https://github.com/biswaroop1547/microbatcher","last_synced_at":"2025-07-06T18:33:34.483Z","repository":{"id":48153883,"uuid":"516622949","full_name":"biswaroop1547/microbatcher","owner":"biswaroop1547","description":"🚀  X-Model server for boosting ML inference without letting you do any heavy-lifting in the backend.","archived":false,"fork":false,"pushed_at":"2022-12-27T21:28:40.000Z","size":74,"stargazers_count":0,"open_issues_count":0,"forks_count":0,"subscribers_count":1,"default_branch":"main","last_synced_at":"2025-04-23T16:38:43.294Z","etag":null,"topics":["api-server","batching","inference","ml","mlops","restful"],"latest_commit_sha":null,"homepage":"","language":"Python","has_issues":true,"has_wiki":null,"has_pages":null,"mirror_url":null,"source_name":null,"license":"mit","status":null,"scm":"git","pull_requests_enabled":true,"icon_url":"https://github.com/biswaroop1547.png","metadata":{"files":{"readme":"README.md","changelog":"CHANGELOG.md","contributing":"CONTRIBUTING.md","funding":null,"license":"LICENSE","code_of_conduct":null,"threat_model":null,"audit":null,"citation":null,"codeowners":null,"security":null,"support":null}},"created_at":"2022-07-22T05:34:33.000Z","updated_at":"2022-12-27T21:33:54.000Z","dependencies_parsed_at":"2023-01-31T05:16:13.847Z","dependency_job_id":null,"html_url":"https://github.com/biswaroop1547/microbatcher","commit_stats":null,"previous_names":[],"tags_count":0,"template":false,"template_full_name":null,"repository_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/biswaroop1547%2Fmicrobatcher","tags_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/biswaroop1547%2Fmicrobatcher/tags","releases_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/biswaroop1547%2Fmicrobatcher/releases","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/biswaroop1547%2Fmicrobatcher/manifests","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/biswaroop1547","download_url":"https://codeload.github.com/biswaroop1547/microbatcher/tar.gz/refs/heads/main","host":{"name":"GitHub","url":"https://github.com","kind":"github","repositories_count":250471120,"owners_count":21435897,"icon_url":"https://github.com/github.png","version":null,"created_at":"2022-05-30T11:31:42.601Z","updated_at":"2022-07-04T15:15:14.044Z","host_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories","repository_names_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repository_names","owners_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners"}},"keywords":["api-server","batching","inference","ml","mlops","restful"],"created_at":"2025-04-23T16:38:45.458Z","updated_at":"2025-04-23T16:38:50.573Z","avatar_url":"https://github.com/biswaroop1547.png","language":"Python","readme":"# \u003cp align=\"center\"\u003eMicroBatcher\u003c/p\u003e\n\u003cp align=\"center\"\u003e\u003cb\u003eX-Model server for boosting ML inference\u003c/b\u003e\u003c/p\u003e\n\u003cp align=\"center\"\u003e\n\n## What is this?\nA server taking care of all your inference to deployment needs + boosting performance without letting you do any heavylifting in the backend.\n\n## Quickstart\n1. Clone this repo \u0026 install\n\n    ```bash\n    git clone https://github.com/biswaroop1547/microbatcher.git \u0026\u0026 cd microbatcher\n    make install\n    ```\n2. Define model path and start server\n    ```bash\n    echo\n    ```\n## Philosophy\nlorem ipsum\n### Why do you need this?\nlorem ipsum\n\n### What will it enable?\nlorem ipsum\n\n#### Features\n* lorem ipsum\n* lorem ipsum\n","funding_links":[],"categories":[],"sub_categories":[],"project_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fbiswaroop1547%2Fmicrobatcher","html_url":"https://awesome.ecosyste.ms/projects/github.com%2Fbiswaroop1547%2Fmicrobatcher","lists_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fbiswaroop1547%2Fmicrobatcher/lists"}