wdndev/llm_interview_note

主要记录大语言大模型(LLMs) 算法(应用)工程师相关的知识及面试题

40
/ 100
Emerging

Covers Transformer architecture fundamentals (attention mechanisms, positional encoding, tokenization) alongside practical implementations like LLaMA and ChatGLM model internals, with dedicated sections on distributed training strategies (data/tensor/pipeline parallelism), inference optimization via vLLM and quantization, and alignment techniques including RLHF and DPO. Integrates with frameworks like DeepSpeed and Megatron, while companion projects (tiny-llm-zh, tiny-rag, tiny-mcp) provide hands-on implementations for pretraining, RAG systems, and MCP-based agents on resource-constrained hardware.

13,130 stars. No commits in the last 6 months.

No License Stale 6m No Package No Dependents
Maintenance 2 / 25
Adoption 10 / 25
Maturity 8 / 25
Community 20 / 25

How are scores calculated?

Stars

13,130

Forks

1,304

Language

HTML

License

Last pushed

Apr 30, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/wdndev/llm_interview_note"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.