https://github.com/cydanix/whisperclip
Privacy-First Voice-to-Text with AI Enhancement for macOS
https://github.com/cydanix/whisperclip
llm-inference local macos macos-app macos-application macosx privacy speech-to-text voice-to-text voice-to-text-transcription
Last synced: 6 days ago
JSON representation
Privacy-First Voice-to-Text with AI Enhancement for macOS
- Host: GitHub
- URL: https://github.com/cydanix/whisperclip
- Owner: cydanix
- License: mit
- Created: 2025-09-27T11:54:02.000Z (12 days ago)
- Default Branch: main
- Last Pushed: 2025-09-27T14:13:58.000Z (12 days ago)
- Last Synced: 2025-09-27T15:05:55.211Z (12 days ago)
- Topics: llm-inference, local, macos, macos-app, macos-application, macosx, privacy, speech-to-text, voice-to-text, voice-to-text-transcription
- Language: Swift
- Homepage: https://whisperclip.com
- Size: 3.68 MB
- Stars: 3
- Watchers: 0
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
# WhisperClip

**Privacy-First Voice-to-Text with AI Enhancement for macOS**
[](https://opensource.org/licenses/MIT)
[](https://www.apple.com/macos/)
[](https://swift.org)[Website](https://whisperclip.com) • [Download](https://whisperclip.com/v2/download) • [Support](mailto:support@cydanix.com)
## ✨ Features
### 🎤 **Voice-to-Text Transcription**
- High-quality speech recognition using WhisperKit
- Multiple model sizes (216MB to 955MB) for different accuracy/speed trade-offs
- Support for multiple languages with auto-detection
- Real-time waveform visualization during recording### 🤖 **AI-Powered Text Enhancement**
- Local LLM processing for grammar correction and text improvement
- Multiple AI models including Gemma, Llama, Qwen, and Mistral
- Custom prompts for different use cases:
- Grammar fixing and email formatting
- Language translation
- Custom text processing workflows### 🔒 **Privacy-First Design**
- **100% local processing** - your voice never leaves your device
- No cloud services, no data collection
- Open source - audit the code yourself
- Secure sandboxed environment### ⚡ **Productivity Features**
- Global hotkey support (Option+Space by default)
- Auto-copy to clipboard
- Auto-paste functionality
- Auto-enter for instant message sending
- Menu bar integration
- Auto-stop recording after 10 minutes### 🎨 **User Experience**
- Beautiful dark-themed interface
- Real-time recording visualization
- Comprehensive onboarding guide
- Easy model management and downloads
- Customizable shortcuts and prompts## 📋 Requirements
- **macOS 14.0** or later
- **20GB** free disk space (for AI models)
- **Microphone access** permission
- **Accessibility permissions** (for global hotkeys)
- **Apple Events permissions** (for clipboard operations)## 🚀 Installation
### Download Pre-built App
1. Visit [whisperclip.com](https://whisperclip.com)
2. Download the latest release
3. Drag WhisperClip.app to your Applications folder
4. Follow the setup guide for permissions### Build from Source
```bash
# Clone the repository
git clone https://github.com/cydanix/whisperclip.git
cd whisperclip# Build the app
./build.sh# For development
./local_build.sh Debug
./local_run.sh Debug
```## 🔧 Usage
### Quick Start
1. **Launch WhisperClip** from Applications or menu bar
2. **Grant permissions** when prompted (microphone, accessibility)
3. **Download AI models** through the setup guide
4. **Press Option+Space** (or click Record) to start recording
5. **Press again to stop** - text will be automatically copied to clipboard### Customization
- **Change hotkey**: Settings → Hotkey preferences
- **Add custom prompts**: Settings → Prompts → Add new prompt
- **Switch AI models**: Setup Guide → Download different models
- **Configure auto-actions**: Settings → Enable auto-paste/auto-enter## 🤖 Supported AI Models
### Speech-to-Text (WhisperKit)
- **OpenAI Whisper Small** (216MB) - Fast, good quality
- **OpenAI Whisper Large v3 Turbo** (632MB) - Best balance
- **Distil Whisper Large v3 Turbo** (600MB) - Optimized speed
- **OpenAI Whisper Large v2 Turbo** (955MB) - Maximum accuracy### Text Enhancement (Local LLMs)
- **Gemma 2 (2B/9B)** - Google's efficient models
- **Llama 3/3.2 (3B/8B)** - Meta's powerful models
- **Qwen 2.5/3 (1.5B-8B)** - Alibaba's multilingual models
- **Mistral 7B** - High-quality French company model
- **Phi 3.5 Mini** - Microsoft's compact model
- **DeepSeek R1** - Advanced reasoning modelAll models run locally using MLX for Apple Silicon optimization.
## 🔒 Privacy & Security
WhisperClip is designed with privacy as the cornerstone:
- **Local Processing Only**: All voice recognition and AI processing happens on your device
- **No Network Requests**: Except for downloading models from Hugging Face
- **No Analytics**: No usage tracking, no telemetry, no data collection
- **Open Source**: Full transparency - inspect the code yourself
- **Sandboxed**: Runs in Apple's secure app sandbox
- **Encrypted Storage**: AI models stored securely on device## 🛠 Development
### Project Structure
```
Sources/
├── WhisperClip.swift # Main app entry point
├── ContentView.swift # Main UI interface
├── AudioRecorder.swift # Voice recording logic
├── VoiceToText*.swift # Transcription engine
├── LLM*.swift # AI text enhancement
├── ModelStorage.swift # Model management
├── SettingsStore.swift # User preferences
└── HotkeyManager.swift # Global shortcuts
```### Dependencies
- **WhisperKit**: Apple's optimized Whisper implementation
- **MLX**: Apple Silicon ML framework
- **MLX-Swift-Examples**: LLM implementations
- **Hub**: Hugging Face model downloads### Building
```bash
# Debug build
./local_build.sh Debug# Release build with code signing
./build.sh# Notarization (requires Apple Developer account)
./notarize.sh
```## 🤝 Contributing
We welcome contributions! Please see our contributing guidelines:
1. **Fork the repository**
2. **Create a feature branch**: `git checkout -b feature/amazing-feature`
3. **Make your changes** and add tests
4. **Commit your changes**: `git commit -m 'Add amazing feature'`
5. **Push to branch**: `git push origin feature/amazing-feature`
6. **Open a Pull Request**### Areas for Contribution
- New AI model integrations
- UI/UX improvements
- Performance optimizations
- Language support
- Accessibility features
- Documentation improvements## 📄 License
WhisperClip is licensed under the **MIT License** - see the [LICENSE](LICENSE) file for details.
This means you can:
- ✅ Use commercially
- ✅ Modify and distribute
- ✅ Use privately
- ✅ Fork and create derivatives**Attribution required**: Please include the original license notice.
## 🏢 About
WhisperClip is developed by **Cydanix LLC**.
- **Website**: [whisperclip.com](https://whisperclip.com)
- **Support**: [support@cydanix.com](mailto:support@cydanix.com)
- **Version**: 1.0.43## 🙏 Acknowledgments
- **Apple** - WhisperKit and MLX frameworks
- **OpenAI** - Original Whisper models
- **Hugging Face** - Model hosting and Hub library
- **ML Community** - Open source AI models (Gemma, Llama, Qwen, etc.)---
**Made with ❤️ for privacy-conscious users**
[⭐ Star this repo](https://github.com/cydanix/whisperclip) if you find it useful!