jsksxs360/How-to-use-Transformers
Transformers 库快速入门教程
Covers core NLP tasks through modular, runnable examples including sequence labeling, machine translation, summarization, and extractive QA, with implementations built on the Hugging Face Transformers library's pipeline and fine-tuning APIs. Structured in four progressive sections from foundational concepts (attention mechanisms, tokenization) through practical applications to large language model training and instruction tuning techniques.
1,850 stars.
Stars
1,850
Forks
223
Language
Python
License
Apache-2.0
Category
Last pushed
Feb 24, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/jsksxs360/How-to-use-Transformers"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related models
ThilinaRajapakse/simpletransformers
Transformers for Information Retrieval, Text Classification, NER, QA, Language Modelling,...
explosion/curated-transformers
🤖 A PyTorch library of curated Transformer models and their composable components
google/deepconsensus
DeepConsensus uses gap-aware sequence transformers to correct errors in Pacific Biosciences...
ruanchaves/hashformers
Accurate word segmentation for hashtags and text, powered by Transformers and Beam Search. A...
EricFillion/happy-transformer
Happy Transformer makes it easy to fine-tune and perform inference with NLP Transformer models.