aaaastark/NBART-Multilingual-Translator
This repository contains a Python script that uses a pre-trained NBART (Neural Bidirectional AutoRegressive Transformer) model to perform multi-lingual translation tasks between several languages. The model was trained on multiple language pairs using data parallelism, allowing it to learn representations across all languages simultaneously.
No commits in the last 6 months.
Stars
1
Forks
—
Language
Jupyter Notebook
License
BSD-3-Clause
Category
Last pushed
Mar 08, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/aaaastark/NBART-Multilingual-Translator"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
Lightning-AI/litgpt
20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.
liangyuwang/Tiny-DeepSpeed
Tiny-DeepSpeed, a minimalistic re-implementation of the DeepSpeed library
catherinesyeh/attention-viz
Visualizing query-key interactions in language + vision transformers (VIS 2023)
microsoft/Text2Grad
🚀 Text2Grad: Converting natural language feedback into gradient signals for precise model...
FareedKhan-dev/Building-llama3-from-scratch
LLaMA 3 is one of the most promising open-source model after Mistral, we will recreate it's...