daekeun-ml/genai-ko-LLM

This hands-on lab walks you through a step-by-step approach to efficiently serving and fine-tuning large-scale Korean models on AWS infrastructure.

33
/ 100
Emerging

Covers QLoRA parameter-efficient fine-tuning using SageMaker training instances and multiple model serving approaches including DeepSpeed, TGI (Text Generation Inference), and NVIDIA FasterTransformer for distributed inference. Integrates with HuggingFace Hub and SageMaker's DJL container, supporting various Korean models (KULLM-Polyglot, KoAlpaca) with notebooks for local debugging before production deployment. Includes RAG implementation examples alongside inference optimization techniques.

No commits in the last 6 months.

Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 7 / 25
Maturity 9 / 25
Community 17 / 25

How are scores calculated?

Stars

26

Forks

8

Language

Jupyter Notebook

License

MIT

Category

llm-fine-tuning

Last pushed

Feb 08, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/generative-ai/daekeun-ml/genai-ko-LLM"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.