FluidInference/FluidAudio

Frontier CoreML audio models in your apps — text-to-speech, speech-to-text, voice activity detection, and speaker diarization. In Swift, powered by SOTA open source.

71
/ 100
Verified

This project helps Apple app developers integrate advanced audio AI features directly into their macOS and iOS applications. It takes raw audio input and can output transcribed text, detect voice activity, identify different speakers, or convert text into spoken audio, all running efficiently on the device itself. App developers can use this to add robust voice capabilities to their products.

1,689 stars. Actively maintained with 98 commits in the last 30 days.

Use this if you are an Apple app developer looking to add fast, private, and on-device speech-to-text, text-to-speech, or speaker recognition features to your macOS or iOS application.

Not ideal if you need a cloud-based audio processing solution or are developing for platforms other than Apple devices.

iOS-development macOS-development speech-recognition voice-user-interface audio-processing
No Package No Dependents
Maintenance 25 / 25
Adoption 10 / 25
Maturity 15 / 25
Community 21 / 25

How are scores calculated?

Stars

1,689

Forks

214

Language

Swift

License

Apache-2.0

Last pushed

Mar 18, 2026

Commits (30d)

98

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/voice-ai/FluidInference/FluidAudio"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.