OpenGVLab/Instruct2Act

Instruct2Act: Mapping Multi-modality Instructions to Robotic Actions with Large Language Model

23
/ 100
Experimental

Uses an LLM to generate executable Python code that orchestrates perception, planning, and control loops for robotic manipulation. The perception pipeline chains foundation models—SAM for object segmentation and CLIP for classification—via predefined APIs that the LLM can invoke, enabling zero-shot task execution without learning-based policies. Integrates with VIMABench for tabletop manipulation evaluation and supports both task-specific and task-agnostic prompting strategies, with optional pointing-language augmentation for object selection.

373 stars. No commits in the last 6 months.

No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 1 / 25
Community 12 / 25

How are scores calculated?

Stars

373

Forks

22

Language

Python

License

Last pushed

Jun 23, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/OpenGVLab/Instruct2Act"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.