Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add new features and fixes to Sparse Update #47

Open
wants to merge 95 commits into
base: main
Choose a base branch
from

Conversation

dnadalini
Copy link
Collaborator

@dnadalini dnadalini commented Jul 25, 2024

This PR adds the following:

  • Leaky Relu Activation Function and related integration in TrainLib_Deployer
  • Integration of Sigmoid Activation Function in TrainLib_Deployer
  • Pseudo-random number generator for Dropout layer
  • L1Loss implementation
  • berHu loss implementation (missing test, to do)

Missing:

  • Integration of pseudo-random number generation in TrainLib_Deployer
  • Integration of L1Loss in TrainLib_Deployer

dnadalini and others added 30 commits April 3, 2024 11:21
Bring changes from main repo
Merge changes from main repo
Rebase main branch to updated main of original repo
Update main branch to original repo
alberto and others added 30 commits August 1, 2024 11:59
…ar layer, that generates qkv from the input, with 3 smaller ones, as the original implementation is designed (making it compatible with publicly available pretrained models). Ran the Black code formatter on the 2 modified py files
…ar layer, that generates qkv from the input, with 3 smaller ones, as the original implementation is designed (making it compatible with publicly available pretrained models). Ran the Black code formatter on the 2 modified py files. Other small fixes
…. Changed some variable names of mhsa fp32 to match official ViT implementation. Other fixes
…e debugged) and files to support the fp16 version
… (FP32, all, with biases, stride and padding)
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants