2025 07 10 google ai future fund
Google AI Future Fund
dwani.ai now has Multi-modal Inference API capability. Along with the Android App, we provide self-hosting options using Open Weight models.
-
Funding - Yes, I am looking for seed grants upto USD 100k for developer salary and organisation expenses for 1 year. dwani.ai is bootstrapped and I have been working full-time since February 2025.
- We require GPU's for dwani.ai inference - we are constrained by GPU availability.
-
Traction - Android App has 300 total installs, with 100+ active installs.
- We have conducted 3 workshops for students. Total 500 students were provided an introduction to get started building AI applications for Indian languages. https://dwani.ai/dwani-ai-workshop.pdf .
- We conduct one workshop per month at Universities. We do not charge the students/university , Entry is open to all. For July we have 2 workshops planned.
-
AI Models and Technology
- https://github.com/dwani-ai/deploy - contains the detailed steps for self-hosting dwani.ai.
- We are using Nvidia GH200/ arm64 on lambda.ai to host one instance of dwani.ai
- Models
- Text + Vision : google/gemma-12B-IT,
- Text to Speech ; ai4bharat/IndicF5, onnx-community/Kokora-82M-v1.0-ONNX
- Speech Transcription - ai4bharat/indic-conformer-600m-mulitlingual, Systran/faster-whisper-larger-v3
- Tool Call : qwen/qwen3-32B
- What is working :
- ASR/Speech transcription is great with AI4Bharat models for Indian languages.
- Text to Speech : IndicF5 is good for Indian languages, but it is compute intensive.
- Gemma3 works well for Text and Vision inference for English. Kannada and German.
- What is on your wishlist ?
- We would like tool_call to be improved in the next release of Gemma,
- I am yet to explore in depth Gemma-3n with Matroshka Transformers for Multi-modal inputs. It had a few issues on release day .