A modular Swift SDK for audio processing with MLX on Apple Silicon
-
Updated
Mar 21, 2026 - Swift
A modular Swift SDK for audio processing with MLX on Apple Silicon
💬 Fast, cross-platform CLI and GUI for batch transcription, translation, speaker annotation and subtitle generation using OpenAI’s Whisper on CPU, Nvidia GPU and Apple MLX.
这是一个基于 mlx-audio 的本地 REST 服务,用来实现兼容 OpenAI 的 TTS / STT 音频接口桥接层。
A high-performance, fully local real-time voice translation agent built for Apple Silicon. Features seamless English-Hindi translation, zero-shot voice cloning, and a stateful agentic workflow orchestrated by LangGraph and MLX-Audio.
A state-of-the-art Web UI for Qwen3-TTS providing zero-shot voice synthesis, optimized natively for Apple Silicon (MLX) and Nvidia (CUDA) with PyTorch fallback integrations.
PageMatch transcribes your audiobook once using NVIDIA's Parakeet model running locally on your Apple Silicon GPU via MLX. After that, finding any moment in a 20-hour book takes under a second — just paste a sentence from the text.
OpenClaw MLX Audio plugin for local TTS on Apple Silicon
Add a description, image, and links to the mlx-audio topic page so that developers can more easily learn about it.
To associate your repository with the mlx-audio topic, visit your repo's landing page and select "manage topics."