I open-sourced the weights, and documented the whole journey, from Torch model conversion, data preparation, JAX training code and inference pipeline. I spent about $300 in egress costs, but it can be as cheap as $100 to train this model (I ran the data collection pipeline on my 5090 Desktop PC as well as fine-tuning Whisper).
Hopefully it can serve as a guide for others that are curious to train these models for other languages (without burning through all the credits trying to fix the pipeline).