zeta
pulsar
Initializing search
kyegomez/zeta
Home
zeta
kyegomez/zeta
Home
Home
Overview
Contributing
Overview
zeta.nn
zeta.nn
zeta.nn.biases
zeta.nn.biases
Xpos
RelativePositionBias
AlibiPositionalBias
DynamicPositionBias
zeta.nn.embeddings
zeta.nn.embeddings
MultiWay
RotaryEmbeddings
TruncatedRotaryEmbedding
PositionalEmbedding
XPOS
YarnEmbedding
VisionEmbedding
SinusoidalEmbeddings
PatchEmbeddings
PositionInterpolationEmbeddings
zeta.nn.modules
zeta.nn.modules
custom_mlp
mbconv
dynamicroutingblock
clippedgeluactivation
mambablock
vittransformerblock
fuseddensegeludense
pscan
adaptive
filmconditioning
mmfusionffn
quickgeluactivation
gatedresidualblock
highwaylayer
multimodalmambablock
rms_norm
ssm
dualpathblock
topngating
mmlayernorm
mm_adapter
laplaceactivation
nfnstem
laser
denseblock
depthwiseconv2d
lora
vlayernorm
flexiconv
pulsar
pool
time_up_sample
spatial_downsample
parallel
conv2dfeedforward
video_autoencoder
recursiveblock
relusquaredactivation
fastgeluactivation
token_learner
layernorm
averagemodelmerger
linearactivation
stochdepth
expert
siglip
ether
newgeluactivation
pytorchgelutanh
multiscaleblock
umambablock
film
adaptive_conv
fused_dropout_layernorm
accurategeluactivation
exo
polymorphic_activation
fusedprojsoftmax
quantizedln
postnorm
moerouter
geluactivation
visionattention
fused_gelu_dense
feedforward
wsconv2d
mlp
slerpmodelmerger
fuseddropoutlayernorm
tripleskipblock
dm
feedbackblock
mixtureofexperts
mamba
perceiverlayer
mishactivation
hebbian
simple_feedback
visual_expert
stochasticskipblock
unet
zeta.nn.attention
zeta.nn.attention
FlashAttention
MultiQueryAttention
MultiheadAttention
FlashAttentionTwo
BaseAttention
LocalAttention
LocalMHA
MixtureOfAttention
MixtureOfAutoregressiveAttention
SparseAttention
zeta.tokenizers
zeta.tokenizers
Language
Language
LanguageTokenizerGPTX
SentencePieceTokenizer
TokenMonster
MultiModal
MultiModal
MultiModalTokenizer
zeta.utils
zeta.utils
Misc
Misc
cast_tuple
group_by_key_prefix
eval_decorator
print_cuda_memory_usage
once
default
gumbel_noise
pad_at_dim
init_zero_
top_p
cast_if_src_dtype
disable_warnings_and_logs
save_load_wrapper
get_sinusoid_encoding_table
main
string_begins_with
gif_to_tensor
l2norm
save_load
log
module_device
print_num_params
top_a
interpolate_pos_encoding_2d
exists
cosine_beta_schedule
track_cuda_memory
maybe
save_memory_snapshot
top_k
print_main
pick_and_pop
track_cuda_memory_usage
group_dict_by_key
video_tensor_to_gift
zeta.ops
zeta.ops
Misc
Misc
img_compose_decompose
img_transpose_2daxis
img_transpose
img_order_of_axes
mos
merge_small_dims
multi_dim_cat
img_compose_bw
squeeze_2d_new
temp_softmax
gumbelmax
_matrix_inverse_root_newton
compute_matrix_root_inverse_residuals
matrix_root_diagonal
sparse_softmax
reshape_audio_to_text
local_softmax
softmaxes
_matrix_root_eigen
main
norm_exp_softmax
multi_dim_split
img_width_to_height
fast_softmax
standard_softmax
unitwise_norm
reshape_video_to_text
img_decompose
unsqueeze_2d_new
reshape_img_to_text
channel_shuffle_new
matrix_inverse_root
sparsemax
gram_matrix_new
logit_scaled_softmax
selu_softmax
reshape_text_to_img
zeta.optim
zeta.optim
Optimizers
Optimizers
StableAdamWUnfused
GradientAscent
DecoupledLionW
SophiaG
zeta.training
zeta.training
Training
Training
fsdp
ParallelWrapper
train
zeta.models
zeta.models
Language and MultiModal
Language and MultiModal
vit
gpt4multimodal
maxvit
llama2
gpt4
andromeda
basemodel
palme
megavit
navit
zeta.structs
zeta.structs
Structures
Structures
Decoder
Transformer
paralleltransformerblock
zeta.quant
zeta.quant
Quantization Algorithms
Quantization Algorithms
QUIK
BitLinear
niva
zeta.rl
zeta.rl
DPO
pulsar
Back to top