{"id":13578326,"url":"https://github.com/alphacep/vosk-api","last_synced_at":"2025-05-12T20:39:22.224Z","repository":{"id":37269804,"uuid":"206138137","full_name":"alphacep/vosk-api","owner":"alphacep","description":"Offline speech recognition API for Android, iOS, Raspberry Pi and servers with Python, Java, C# and Node","archived":false,"fork":false,"pushed_at":"2025-05-01T08:58:24.000Z","size":14420,"stargazers_count":9406,"open_issues_count":525,"forks_count":1266,"subscribers_count":124,"default_branch":"master","last_synced_at":"2025-05-05T16:04:15.561Z","etag":null,"topics":["android","asr","deep-learning","deep-neural-networks","deepspeech","google-speech-to-text","ios","kaldi","offline","privacy","python","raspberry-pi","speaker-identification","speaker-verification","speech-recognition","speech-to-text","speech-to-text-android","stt","voice-recognition","vosk"],"latest_commit_sha":null,"homepage":"","language":"Jupyter Notebook","has_issues":true,"has_wiki":null,"has_pages":null,"mirror_url":null,"source_name":null,"license":"apache-2.0","status":null,"scm":"git","pull_requests_enabled":true,"icon_url":"https://github.com/alphacep.png","metadata":{"files":{"readme":"README.md","changelog":null,"contributing":null,"funding":null,"license":"COPYING","code_of_conduct":null,"threat_model":null,"audit":null,"citation":null,"codeowners":null,"security":null,"support":null,"governance":null,"roadmap":null,"authors":null,"dei":null,"publiccode":null,"codemeta":null,"zenodo":null}},"created_at":"2019-09-03T17:48:42.000Z","updated_at":"2025-05-05T15:01:15.000Z","dependencies_parsed_at":"2023-01-31T08:00:58.625Z","dependency_job_id":"7f326188-63f2-4b40-af31-22fda4914520","html_url":"https://github.com/alphacep/vosk-api","commit_stats":{"total_commits":494,"total_committers":41,"mean_commits":"12.048780487804878","dds":"0.13967611336032393","last_synced_commit":"1b308a30178aa032b32884df29110d0c007b791c"},"previous_names":[],"tags_count":30,"template":false,"template_full_name":null,"repository_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/alphacep%2Fvosk-api","tags_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/alphacep%2Fvosk-api/tags","releases_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/alphacep%2Fvosk-api/releases","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/alphacep%2Fvosk-api/manifests","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/alphacep","download_url":"https://codeload.github.com/alphacep/vosk-api/tar.gz/refs/heads/master","host":{"name":"GitHub","url":"https://github.com","kind":"github","repositories_count":253817643,"owners_count":21969016,"icon_url":"https://github.com/github.png","version":null,"created_at":"2022-05-30T11:31:42.601Z","updated_at":"2022-07-04T15:15:14.044Z","host_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories","repository_names_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repository_names","owners_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners"}},"keywords":["android","asr","deep-learning","deep-neural-networks","deepspeech","google-speech-to-text","ios","kaldi","offline","privacy","python","raspberry-pi","speaker-identification","speaker-verification","speech-recognition","speech-to-text","speech-to-text-android","stt","voice-recognition","vosk"],"created_at":"2024-08-01T15:01:29.467Z","updated_at":"2025-05-12T20:39:22.189Z","avatar_url":"https://github.com/alphacep.png","language":"Jupyter Notebook","readme":"# Vosk Speech Recognition Toolkit\n\nVosk is an offline open source speech recognition toolkit. It enables\nspeech recognition for 20+ languages and dialects - English, Indian\nEnglish, German, French, Spanish, Portuguese, Chinese, Russian, Turkish,\nVietnamese, Italian, Dutch, Catalan, Arabic, Greek, Farsi, Filipino,\nUkrainian, Kazakh, Swedish, Japanese, Esperanto, Hindi, Czech, Polish.\nMore to come.\n\nVosk models are small (50 Mb) but provide continuous large vocabulary\ntranscription, zero-latency response with streaming API, reconfigurable\nvocabulary and speaker identification.\n\nSpeech recognition bindings implemented for various programming languages\nlike Python, Java, Node.JS, C#, C++, Rust, Go and others.\n\nVosk supplies speech recognition for chatbots, smart home appliances,\nvirtual assistants. It can also create subtitles for movies,\ntranscription for lectures and interviews.\n\nVosk scales from small devices like Raspberry Pi or Android smartphone to\nbig clusters.\n\n# Documentation\n\nFor installation instructions, examples and documentation visit [Vosk\nWebsite](https://alphacephei.com/vosk).\n","funding_links":[],"categories":["Jupyter Notebook","Install from Source","C++","语音识别","Repos","Related Projects","人工智能","Speech Processing"],"sub_categories":["Text-To-Speech Synthesis (TTS)","网络服务_其他","[Tools](#tools-1)","Speech-to-Text"],"project_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Falphacep%2Fvosk-api","html_url":"https://awesome.ecosyste.ms/projects/github.com%2Falphacep%2Fvosk-api","lists_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Falphacep%2Fvosk-api/lists"}