Winky is a cross-platform desktop voice assistant that helps you quickly convert voice to text and run intelligent LLM-powered actions. With a convenient floating microphone overlay, you can interact with Winky from anywhere on your screen, making it perfect for productivity workflows.
Winky supports both cloud-based and local AI processing, giving you the flexibility to choose between speed and privacy. Whether you need quick voice commands, transcription, or AI-powered responses, Winky is ready to help.
- Extension repository: https://github.com/Artasov/winky-ext
- Chrome Web Store: https://chromewebstore.google.com/detail/winky/mpinlhhkmpljjlcekiocnglfbfpamkjl
This repository contains the source code for Winky, a cross-platform desktop application built with Tauri + React + Vite. The application provides a modern, efficient voice assistant experience with support for multiple AI providers and local processing options.
- FREE USAGE - no subscription required, no limits for local processing
- Voice Recognition - advanced speech-to-text conversion with multiple AI models
- LLM Processing - intelligent AI-powered actions and responses
- Floating Microphone - convenient floating microphone overlay for quick access
- Local Speech Recognition - use local AI models for faster processing and privacy
- Quick Actions - customizable hotkeys and actions for productivity
- Privacy & Security - all data processed locally, audio is not stored
- Cross-platform - works on Windows, macOS and Linux
- Simple interface - intuitive and easy to use
- Customizable - configure transcription models, LLM providers, and actions
If you have any issues using the app, please open an issue
- Open
Winkyapplication - Complete the initial setup wizard:
- Sign in with your account (OAuth authentication)
- Configure your API keys:
OpenAI API key(get it from platform.openai.com)Google AI API key(get it from console.cloud.google.com)
- Choose your speech recognition mode:
Cloud- use cloud-based transcription (OpenAI Whisper, Google AI)Local- use local fast-whisper for privacy and speed
- Configure LLM settings:
- Choose your preferred LLM provider
- Select the model suitable for your needs
- Set up quick actions:
- Configure custom hotkeys for actions
- Create and customize your action workflows
- Use the floating microphone overlay to start voice recognition
- Speak your command or question
- Get instant AI-powered responses and actions
- Use hotkeys for quick access to common actions
- Access your profile, actions, and settings from the main window
- Position the floating microphone overlay where it's convenient for you
- Customize hotkeys to match your workflow
- Use local speech recognition for better privacy
- Practice with different commands to get the best results
The examples below are implemented and tested on Windows 11. Steps may differ on other systems.
-
In
Winkysettings selectMode -> Speech Recognition=Local. -
In
Winkysettings choose one ofModel -> Speech Recognition -
In
Winkysettings chooseLocal transcription device:GPU(Graphics/NVIDIA) orCPU(Processor)
The local speech recognition server will be automatically installed and managed by Winky.
Minimum recommended configuration:
- CPU - 4 cores / 8 threads
- GPU - 6 GB VRAM
- RAM - 16 GB
-
In
Winkysettings selectMode -> LLM=Local. -
In
Winkysettings choose aModel -> LLMfrom the available models (Ollama models) -
ollama pull <model-name>
-
ollama serve
The first use after the opening of the program will be slower, since with local use of the AI models will be loaded in GPU or RAM, which takes time. Before important tasks, do a test run so that the subsequent calls are faster.
We welcome contributions to the project! If you want to contribute:
- Fork the repository
- Create a feature branch (
git checkout -b feature/amazing-feature) - Commit your changes (
git commit -m 'Add amazing feature') - Push to the branch (
git push origin feature/amazing-feature) - Open a Pull Request
See CONTRIBUTING.md for detailed guidelines.
- Node.js 20+ (LTS)
- Rust 1.80+ (for building Tauri)
- npm or yarn
# Clone the repository
git clone https://github.com/placeholder/winky.git
cd winky
# Install dependencies
npm install
# Build the project
npm run build
# Run in development mode
npm run devsrc/
βββ renderer/ # React renderer process (UI)
β βββ app/ # Application logic and hooks
β βββ components/ # React components
β βββ context/ # React context providers
β βββ features/ # Feature modules
β βββ services/ # API and service layer
β βββ windows/ # Window components
β βββ ...
βββ shared/ # Shared types and utilities
βββ ...
src-tauri/
βββ src/ # Rust backend (Tauri)
βββ ...
npm run dev- run in development modenpm run build- build the projectnpm run build:renderer- build only the renderer (frontend)npm run dev:renderer- run renderer dev server onlynpm run lint- check TypeScript typesnpm run typecheck- same as lintnpm run preview- preview built frontend
npm run buildCreates:
- Portable executable in
src-tauri/target/release/
npm run buildCreates:
- DMG archive for Intel and Apple Silicon
Note: For macOS builds, you may need to:
- Install Xcode Command Line Tools:
xcode-select --install
npm run buildCreates:
- Portable directory in
src-tauri/target/release/
- Tauri - cross-platform desktop application framework
- React - UI library
- TypeScript - typed JavaScript
- Tailwind CSS - utility-first CSS framework
- Vite - build tool and dev server
- OpenAI API - AI integration
- Google AI API - AI integration
Made with β€οΈ for productivity and assistance
