Skip to content

Zeta

Build SOTA AI Models 80% faster with modular, high-performance, and scalable building blocks!

Docs

MIT License MIT License

Join our Discord Subscribe on YouTube Connect on LinkedIn Follow on X.com

GitHub issues GitHub forks GitHub stars GitHub licenseGitHub star chartDependency Status Downloads

Join the Agora discordShare on Twitter Share on Facebook Share on LinkedIn

Share on Reddit Share on Hacker News Share on Pinterest Share on WhatsApp

After building out thousands of neural nets and facing the same annoying bottlenecks of chaotic codebases with no modularity and low performance modules, Zeta needed to be born to enable me and others to quickly prototype, train, and optimize the latest SOTA neural nets and deploy them into production.

Zeta places a radical emphasis on useability, modularity, and performance. Zeta is now currently employed in 100s of models across my github and across others. Get started below and LMK if you want my help building any model, I'm here for you 😊 πŸ’œ

Architecture

zeta/
β”œβ”€β”€ experimental/      # Contains experimental features for testing future capabilities
β”œβ”€β”€ models/            # Houses model architectures and neural network definitions
β”œβ”€β”€ nn/                # Core neural network layers and utilities for building models
β”œβ”€β”€ ops/               # Low-level operations and mathematical functions
β”œβ”€β”€ optim/             # Optimization algorithms for training
β”œβ”€β”€ rl/                # Reinforcement learning components and tools
β”œβ”€β”€ structs/           # Data structures and utilities for managing model states
β”œβ”€β”€ tokenizers/        # Tokenization modules for processing data (text, etc.)
β”œβ”€β”€ training/          # High-level training loops and training utilities
β”œβ”€β”€ utils/             # General-purpose utilities and helper functions
└── __init__.py        # Initializes the framework and handles global imports

Zeta’s Abstraction:

The Zeta framework abstracts over PyTorch and CUDA, aiming to provide flexibility in building, training, and deploying models. Each module serves a distinct role, allowing users to construct neural networks, define custom operations, and handle everything from low-level ops to high-level training routines.

This structure provides several key benefits: - Modularity: Each module encapsulates specific functionality, making it easy to extend or modify. - Flexibility: Zeta integrates seamlessly with PyTorch and CUDA, but offers a more structured and organized way to build models. - Performance: By building on CUDA, Zeta ensures efficient computation while maintaining ease of use through PyTorch's high-level abstractions.

This tree structure reflects the framework’s intent to simplify complex deep learning operations while providing the flexibility to customize each layer and operation according to specific use cases.