A small data-collection and sign-language recognition project. This repository contains code to collect gesture data, train a model, and run realtime inference.
Repository layout (important files):
data/- collected .npy feature files (excluded from git by default)config/actions_config.py- action / label configurationmodels/- saved models and encoders (model files are excluded by .gitignore)src/- application codecollect_data.py/collect_data_gui.py- scripts to collect and label datatrain_model.py- training script (produces a .h5 model)realtime_inference.py- script for realtime camera inference
TODO.md- project TODOs
- Create and activate a virtual environment (recommended):
python3 -m venv .venv
source .venv/bin/activate- Install dependencies:
pip install -r requirements.txt- Collect data (example):
python src/collect_data.py
# or use the GUI
python src/collect_data_gui.py- Train model:
python src/train_model.py- Run realtime inference (requires camera and a trained model in
models/):
python src/realtime_inference.py- Large binary files (models, .npy datasets) are ignored in
.gitignore. Store them outside the repo or use Git LFS if you want them in the remote. - If you need a GPU-enabled TensorFlow build, install the appropriate
tensorflowpackage for your platform.
Feel free to open issues or pull requests. If you'd like, I can add CI (GitHub Actions) to run linting and unit tests.
Add a LICENSE file if you want to set a license for this project.