Skip to content

Latest commit

 

History

History
47 lines (41 loc) · 1.43 KB

TODOs.md

File metadata and controls

47 lines (41 loc) · 1.43 KB

Modules to be added.

  • Average Pooling
  • Transformer Decoder Layer
  • Transformer Encoder
  • Transformer Decoder
  • Batch Normalization
  • Group Normalization
  • More activation functions
  • Loss functions in functional module
  • NLLLoss
  • BCELoss
  • CTCLoss
  • Locally Connected modules
  • Multihead Attention with RoPE module
  • Multihead Attention with relative positional embedding module
  • RNN module
  • LSTM module
  • GRU module
  • Hinge Loss
  • Huber Loss
  • KL Divergence
  • Cosine Similarity
  • Poisson Loss
  • RMSprop Optimizer
  • SGD Optimizer
  • Adagrad Optimizer
  • Module Dict
  • ConvBert
  • Serialization
  • Saving & Loading weights
  • Convolution Transpose

You are welcome to contribute by adding any of these missing modules or if there is another module that you want to add you can submit a pull request or an issue.

Note: Before implementing any of the mentioned loss functions, check optax, It's recommended to use it and you might find the loss function implemented and you will just need to wrap it with the hera.losses.Loss class.

Note: Before implementing any of the mentioned optimizers, check optax, It's recommended to use it and you might find the optimizer implemented and you will just need to wrap it with the hera.optimizers.Optimizer class.