line/sacpo

[NeurIPS 2024] SACPO (Stepwise Alignment for Constrained Policy Optimization)

28
/ 100
Experimental

This project helps AI developers and researchers refine large language models (LLMs) to be both helpful and safe. It takes an existing LLM and training datasets focused on helpfulness and safety, then outputs fine-tuned models and evaluations, showing how well the model adheres to both helpfulness and safety constraints. It's ideal for those building conversational AI, chatbots, or assistants where output quality and user safety are critical.

No commits in the last 6 months.

Use this if you are developing or fine-tuning large language models and need to systematically improve their helpfulness while rigorously enforcing safety guidelines.

Not ideal if you are a business user looking for a no-code solution to customize an LLM, as this requires technical expertise in machine learning and Python.

large-language-models ai-safety model-alignment natural-language-processing conversational-ai
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 4 / 25
Maturity 16 / 25
Community 8 / 25

How are scores calculated?

Stars

8

Forks

1

Language

Python

License

Apache-2.0

Last pushed

Dec 23, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/line/sacpo"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.