--- license: mit datasets: - MuzzammilShah/people-names language: - en model_name: Batch Normalization for Neural Networks library_name: pytorch tags: - makemore - batch-normalization - neural-networks - andrej-karpathy --- # Batch Normalization for Neural Networks: Makemore (Part 3) In this repository, I implemented **Batch Normalization** within a neural network framework to enhance training stability and performance, following Andrej Karpathy's approach in the **Makemore - Part 3** video. ## Overview This implementation focuses on: - **Normalizing activations and gradients**. - Addressing initialization issues. - Utilizing Kaiming initialization to prevent saturation of activation functions. Additionally, **visualization graphs** were created at the end to analyze the effects of these techniques on the training process and model performance. ## Documentation For a better reading experience and detailed notes, visit my **[Road to GPT Documentation Site](https://muzzammilshah.github.io/Road-to-GPT/Makemore-part3/)**. ## Acknowledgments Notes and implementations inspired by the **Makemore - Part 3** video by [Andrej Karpathy](https://karpathy.ai/). For more of my projects, visit my [Portfolio Site](https://muhammedshah.com).