Skip to content

Abhishek-4px/VisionTransformer

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

43 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

VisionTransformer

This is implementation of a vision transformer , that basically uses the encoder part only of the Transformer . For more clear unerstanding I have added comments on each step where I thought they are necessary . You can play along with the hyperparameters and also you may consider adding a Learning Rate scheduler for keeping it high at the start and then lowering it to prevent Divergence . You can also increase EPOCHS ( I used free GPU from colab so I could not train it to the end just stopped at 75% , but it was visible that the model was still not platued and converging after 20 Epochs )

This is the complete structure of Transformer but we will be using only the encoder block Replacing the FeedForward Network with a MultiLayerPerceptron
Transfromer

These are the computed accurcies around 20 Epochs Accuracies Plot

These are few visualized examples out of all Results

About

This is implementation of a vision transformer , that basically uses the encoder part only of the Transformer .

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors