Text Generation
Transformers
PyTorch
Safetensors
gpt2
conversational
text-generation-inference

translation

#10
by exoplanet - opened

Hi team,
Would you recommend using this model for translating [ sv da nb ] to/from en ?
Cheers.

AI Sweden Model Hub org

Yeah, try it, even the 6.7b models if you can.

Also check out: https://huggingface.co/AI-Sweden-Models/gpt-sw3-6.7b-v2-translator

Oh, if only my target environment had more memory, that translator fine-tuned model would be awesome, however up to 4B params would fit the bill. As I'm not native a speaker of sv / da / nb, I cannot tell whether the translations are good or somewhat hallucinated. Therefore, should I opt for a smaller translation fine-tuned models instead?

I have tried sv-en to/from using AI-Sweden (both GPT 1.3b and 6.7b v2 translator), X-ALMA and Helsinki-NLP. This model works good enough for translation but the 6.7b v2 Translator is way better in niche situations, but still NOT good on Industry specific language (None of the models are). X-LLama is similar to AI-Sweden/GPT sw3 6.7b v2 translator might be somewhat better but that expected from such a larger model.

The super lightweight one way translators from Helsinki-NLP are better than GPT 1.3b but worse compared to the bigger models. So if your limited in compute space (not storage due to needing 6 models from the Helsinki-NLP for to/from on 3 languages) then its a good choice. https://huggingface.co/Helsinki-NLP

Thanks @Kenopet for these great insights!

exoplanet changed discussion status to closed

Sign up or log in to comment