• Skip to primary navigation
  • Skip to content
  • Skip to footer
Shitanshu Bhushan
  • About
  • Posts
  • Categories
  • Tags
    Shitanshu Bhushan

    Shitanshu Bhushan

    Graduate student at University of Michigan, Ann Arbor

    • Resume
    • Linkedin
    • GitHub

    Linearizing Attention

    less than 1 minute read

    Breaking the quadratic barrier: modern alternatives to softmax attention

    Direct Link

    Tags: Attention, Blog, from scratch, Machine learning

    Categories: Blog

    Updated: December 26, 2024

    Share on

    Twitter Facebook LinkedIn
    Previous Next

    You May Also Enjoy

    The Math Behind In-Context Learning Permalink

    less than 1 minute read

    From attention to gradient descent: unraveling how transformers learn from examples

    Deeper is better: Coding ResNet from scratch Permalink

    less than 1 minute read

    Simple implementation of ResNet-50 from scratch using pytorch

    Random Forest & AdaBoost: Coding Ensemble Methods from Scratch Permalink

    less than 1 minute read

    Simple implementation of Random Forest & AdaBoost(SAMME)

    Decision Trees: Branching Out Step by Step Permalink

    less than 1 minute read

    Step-by-step guide on building a Decision Tree using Gini impurity

    • Linkedin
    • GitHub
    • Feed
    © 2025 Shitanshu Bhushan. Powered by Jekyll & Minimal Mistakes.