daekeun-ml/genai-ko-LLM
This hands-on lab walks you through a step-by-step approach to efficiently serving and fine-tuning large-scale Korean models on AWS infrastructure.
Covers QLoRA parameter-efficient fine-tuning using SageMaker training instances and multiple model serving approaches including DeepSpeed, TGI (Text Generation Inference), and NVIDIA FasterTransformer for distributed inference. Integrates with HuggingFace Hub and SageMaker's DJL container, supporting various Korean models (KULLM-Polyglot, KoAlpaca) with notebooks for local debugging before production deployment. Includes RAG implementation examples alongside inference optimization techniques.
No commits in the last 6 months.
Stars
26
Forks
8
Language
Jupyter Notebook
License
MIT
Category
Last pushed
Feb 08, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/generative-ai/daekeun-ml/genai-ko-LLM"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
GURPREETKAURJETHRA/Llama-3-ORPO-Fine-Tuning
Llama 3 ORPO Fine Tuning on A100 in Colab Pro.
ramalamadingdong/onnx-rubikpi
ONNX LLM runtime on RUBIK-Pi with Gemma 1B and Llama 3.2 1B
keanteng/sesame-csm-elise
Fine-Tuning Sesame CSM Wth Elise. Enjoy the voice ( ̄︶ ̄)↗
sukanyabag/Finetuning-Qwen2-7B-VQA-on-Radiology-Scans
This repository is doing the finetuning of the Qwen2 7B VLM for performing VQA (Visual Question...
DianaDorobantu/legal-llm
Develop a Romanian legal domain Large Language Model (LLM) using pre-trained model and...