tiny-cuda-nn and neural-network-cuda
With tool A being a lightning-fast framework and tool B being a neural network built from scratch, they are **competitors**, as both offer implementations of neural networks using CUDA/C++, but A appears to be a more mature and performant solution.
About tiny-cuda-nn
NVlabs/tiny-cuda-nn
Lightning fast C++/CUDA neural network framework
Provides fully-fused MLP kernels and multiresolution hash grid encodings optimized for neural field applications, with optional JIT compilation that fuses encoding, network, and custom operations into single CUDA kernels for 1.5–5x speedups. Offers a JSON-configurable C++ API supporting various encodings, losses, and optimizers, with Python bindings for PyTorch integration and lower-level CUDA RTC APIs for embedding models directly into application kernels.
About neural-network-cuda
BobMcDear/neural-network-cuda
Neural network from scratch in CUDA/C++
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work