Zeta¶
Build SOTA AI Models 80% faster with modular, high-performance, and scalable building blocks!
After building out thousands of neural nets and facing the same annoying bottlenecks of chaotic codebases with no modularity and low performance modules, Zeta needed to be born to enable me and others to quickly prototype, train, and optimize the latest SOTA neural nets and deploy them into production.
Zeta places a radical emphasis on useability, modularity, and performance. Zeta is now currently employed in 100s of models across my github and across others. Get started below and LMK if you want my help building any model, I'm here for you π π
Architecture¶
zeta/
βββ experimental/ # Contains experimental features for testing future capabilities
βββ models/ # Houses model architectures and neural network definitions
βββ nn/ # Core neural network layers and utilities for building models
βββ ops/ # Low-level operations and mathematical functions
βββ optim/ # Optimization algorithms for training
βββ rl/ # Reinforcement learning components and tools
βββ structs/ # Data structures and utilities for managing model states
βββ tokenizers/ # Tokenization modules for processing data (text, etc.)
βββ training/ # High-level training loops and training utilities
βββ utils/ # General-purpose utilities and helper functions
βββ __init__.py # Initializes the framework and handles global imports
Zetaβs Abstraction:¶
The Zeta framework abstracts over PyTorch and CUDA, aiming to provide flexibility in building, training, and deploying models. Each module serves a distinct role, allowing users to construct neural networks, define custom operations, and handle everything from low-level ops to high-level training routines.
This structure provides several key benefits: - Modularity: Each module encapsulates specific functionality, making it easy to extend or modify. - Flexibility: Zeta integrates seamlessly with PyTorch and CUDA, but offers a more structured and organized way to build models. - Performance: By building on CUDA, Zeta ensures efficient computation while maintaining ease of use through PyTorch's high-level abstractions.
This tree structure reflects the frameworkβs intent to simplify complex deep learning operations while providing the flexibility to customize each layer and operation according to specific use cases.