derek-byte/multilingual-voice-assistant-llm

cohere labs - aya expedition 2025: integrating speech & audio into aya vision (all in one toolkit for integrating tts/stt) ⚙️

30
/ 100
Emerging

This toolkit helps you integrate speech and audio processing into AI systems like Aya Vision. It takes spoken language (audio) and converts it into text, and also converts text into spoken language. This allows AI to understand voice commands and respond verbally, making interactions more natural and accessible. It's designed for AI developers working on multimodal applications.

No commits in the last 6 months.

Use this if you are developing AI applications that need to process voice input (Speech-to-Text) and generate spoken responses (Text-to-Speech) to enhance user interaction and accessibility.

Not ideal if you are looking for a standalone end-user application for speech translation or voice assistance, as this is a toolkit for developers.

AI-development multimodal-AI speech-recognition text-to-speech AI-accessibility
Stale 6m No Package No Dependents
Maintenance 2 / 25
Adoption 4 / 25
Maturity 16 / 25
Community 8 / 25

How are scores calculated?

Stars

8

Forks

1

Language

Python

License

Apache-2.0

Last pushed

May 15, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/voice-ai/derek-byte/multilingual-voice-assistant-llm"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.