Back

Krutrim Translate

Low latency and long context SoTA Text to Text Translation

Text to text translate
~450M params
Overview
Playground

Description

The Krutrim translate model translates the input text into one of the chosen Indic languages. To build Krutrim translate, we increased the context length of the popular IndicTrans2 translation model, extending it from 256 to 4096. For training, we leveraged the Bharat Parallel Corpus Collection (BPCC) while also augmenting it with our own data to enhance performance.

Furthermore, to improve latency, we explored various architectures for both training and distillation. We are open-sourcing the distilled version with 6 encoder and 3 decoder layers, supporting translation in both directions: English to Indic and Indic to English. This architecture achieves at least a 4x reduction in latency compared to both the original IndicTrans2 and the distilled IndicTrans2 models, with minimal decline in performance.

The following is the list of languages supported by our model: English, Bengali, Hindi, Kannada, Marathi, Malayalam, Gujarati, Telugu, and Tamil.

Animated GIF

Use-Cases:

  • 1. Multilingual Communication – Enables seamless conversations between people speaking different Indic languages.
  • 2. Media & Entertainment – Translates movies, TV shows, and podcasts for Indic audiences.
  • 3. Education & E-learning – Helps in translating lectures, training materials, and online courses.
  • 4. Travel & Tourism – Assists travelers with translation of speech.
  • 5. Customer Support – Enhances multilingual customer service through automated translations.
  • 6. Healthcare – Facilitates doctor-patient communication in different languages.
  • 7. Business & Corporate – Enables meetings and document translation for multinational teams.
  • 8. Legal & Government – Supports courtroom interpretation and official document translation.

And many more...

Evaluation Results

As we benchmarked our model against IndicTrans2, we evaluated its performance using the IN22-gen and IN22-conv datasets. The IndicTrans2 results were sourced from their research paper. Below, we present a comparison of the CHRF++ scores achieved by both models.

LanguageIN22-genIN22-conv
Eng → IndInd → EngEng → IndInd → Eng
Bengali51.850.063.260.8
Hindi56.754.465.462.0
Kannada51.047.964.258.4
Marathi51.048.963.760.7
Malayalam50.949.364.557.8
Gujarati53.551.466.557.7
Punjabi50.650.263.458.1
Telugu52.450.064.860.4
Tamil49.548.359.856.9
Average51.950.063.959.2

How to access this model

License

This code repository and the model weights are licensed under the Krutrim Community License.