Human-Activity-Recognition-using-CNN and TensorFlow-on-Android-for-Human-Activity-Recognition-with-LSTMs
About Human-Activity-Recognition-using-CNN
aqibsaeed/Human-Activity-Recognition-using-CNN
Convolutional Neural Network for Human Activity Recognition in Tensorflow
This project helps you classify everyday physical actions, like walking, jogging, or sitting, from sensor data. It takes raw data from smartphone accelerometers and gyroscopes and tells you exactly what activity a person was performing. This tool is useful for researchers or product developers working with wearable technology or health monitoring applications.
About TensorFlow-on-Android-for-Human-Activity-Recognition-with-LSTMs
curiousily/TensorFlow-on-Android-for-Human-Activity-Recognition-with-LSTMs
iPython notebook and Android app that shows how to build LSTM model in TensorFlow and deploy it on Android
This project helps you classify human activities using sensor data from a smartphone. You feed in raw accelerometer and gyroscope readings, and it tells you what activity a person is performing, such as walking or standing. This is useful for mobile app developers or researchers creating fitness trackers, health monitoring apps, or interactive sports applications.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work