LSTM-Human-Activity-Recognition and RNN-for-Human-Activity-Recognition-using-2D-Pose-Input

Maintenance 0/25
Adoption 10/25
Maturity 16/25
Community 25/25
Maintenance 0/25
Adoption 10/25
Maturity 8/25
Community 23/25
Stars: 3,549
Forks: 938
Downloads:
Commits (30d): 0
Language: Jupyter Notebook
License: MIT
Stars: 296
Forks: 76
Downloads:
Commits (30d): 0
Language: Jupyter Notebook
License:
Stale 6m No Package No Dependents
No License Stale 6m No Package No Dependents

About LSTM-Human-Activity-Recognition

guillaume-chevalier/LSTM-Human-Activity-Recognition

Human Activity Recognition example using TensorFlow on smartphone sensors dataset and an LSTM RNN. Classifying the type of movement amongst six activity categories - Guillaume Chevalier

# Technical Summary Employs a many-to-one LSTM architecture that processes 128-sample time windows of 9-channel inertial sensor data (3-axis accelerometer and gyroscope readings) without extensive feature engineering, relying instead on the recurrent network to automatically learn temporal patterns across sequential measurements. Minimal preprocessing is applied beyond gravity filtering, contrasting with traditional signal-processing-heavy approaches that require manual feature extraction. Built with TensorFlow and includes Jupyter notebook implementations demonstrating end-to-end data loading, model training, and evaluation metrics on the UCI HAR Dataset.

About RNN-for-Human-Activity-Recognition-using-2D-Pose-Input

stuarteiffert/RNN-for-Human-Activity-Recognition-using-2D-Pose-Input

Activity Recognition from 2D pose using an LSTM RNN

Scores updated daily from GitHub, PyPI, and npm data. How scores work