Self-Hosted Voice AI Workflow - From voice to action. Powered by on-prem AI.
by•
Meet Devaten— a fully on-premise AI voice automation tool, deployed via Docker Compose. It converts voice to text, runs the text through AI classification, and then notifies users or updates your source database — all without sending data to the cloud.


Replies
Noloco
This is a great idea @mikko_larkela will definitely have to try it out
Do you connect to your own models or does the container ship with the models? (Does this make the containers quite large?)
@domhnall_o_hanlon Glad you liked the idea 😊
the container itself doesn’t ship with all AI models preloaded, though OpenAI Whisper is already included in the base image.
The only additional model we’ve tested with so far is LLaMA 3 (8B). So after the Docker containers are up and running, you just need to run:
ollama pull llama3:8b
This approach helps keep the Docker images lighter (as shown in the docker system df output—total image size is around 12.5 GB), and gives flexibility to pull only the models you need.
Here is guide to try this:
https://github.com/devatengit/onpremises-common-agent/tree/main
Noloco
@mikko_larkela thanks!
@domhnall_o_hanlon
A few things to note:
After the Docker Compose setup is up and running, and Ollama has been pulled,
You can start using the APIs. also.
You can send audio files and receive transcriptions and classifications via AI.
http://host:8081/devaten/audio/transcribe/?file
The current prompts are defined in the docker-compose file. There is a demo set up for "Patients" classification, which you can customize to fit your own use case.
ollama.request.prompt:
ollama.classification.prompt: