Acknowledgements
hama exists thanks to collaborative efforts across research, engineering, and community sponsors. We are grateful for the following contributions.
Core maintainers
Thank you to the maintainers who keep hama open source.
Community
Thank you to everyone who reports issues, shares ideas, and helps steer the project.
Training data
G2P
- English: CMUDict, IPADict, and English Wikipron.
- Korean: Korean Wikipron.
Chinese (Hans/Hant), Dutch, French, German, Italian, Japanese, Portuguese, Spanish, and Thai: Wikipron TSV exports used in the multilingual manifests.
ASR
- LibriSpeech, including the
train-clean-100,train-clean-360, andtrain-other-500training splits. - Mozilla Common Voice English.
- AMI meeting speech via the open
amisplit fromesb/datasets.
Sponsors
- Sunah Hong — donated an NVIDIA DGX Spark for training and benchmarking.
Upstream projects
- ONNX Runtime for powering cross-platform inference.
- Bun, Node, and the Python ecosystem for reliable tooling.
- Astro for the content/workflow experience.
- 나눔 손글씨 김유이체 for the logo wordmark font, courtesy of NAVER CLOVA.
If you or your organization rely on hama, support the project via GitHub Sponsors. For business inquiries, reach out at intelligent.hama@gmail.com.