Dataset Viewer
Auto-converted to Parquet Duplicate
source
stringlengths
65
120
full_text
stringlengths
2.6k
14.4k
title
stringlengths
22
97
https://nepalprabin.github.io./posts/2020-06-05-paper-explanation-going-deeper-with-convolutions-googlenet.html
Paper Explanation: Going deeper with Convolutions (GoogLeNet) Google proposed a deep Convolution Neural Network named inception that achieved top results for classification and detection in ILSVRC 2014. The ImageNet Large Scale Visual Recognition Challenge (ILSVRC) evaluates algorithms for object detection and image ...
Paper Explanation: Going deeper with Convolutions (GoogLeNet)
https://nepalprabin.github.io./posts/2021-07-27-illustrated-vision-transformers.html
Illustrated Vision Transformers Introduction Ever since Transformer was introduced in 2017, there has been a huge success in the field of Natural Language Processing (NLP). Almost all NLP tasks use Transformers and it’s been a huge success. The main reason for the effectiveness of the Transformer was its ability to h...
Illustrated Vision Transformers
https://nepalprabin.github.io./posts/2023-05-15-gpt4-summary.html
Brief overview of GPT-4 Since the release of ChatGPT, there has been significant interest and discussion within the broader AI and natural language processing communities regarding its capabilities. In addition to this, ChatGPT has captured the attention of the internet at large due to its remarkable ability to genera...
Brief overview of GPT-4
https://nepalprabin.github.io./posts/2020-08-23-neural-style-transfer-and-its-working.html
Neural style transfer and its working Have you ever used an app called Prisma that styles your image using popular paintings and turns your photo stunning? If that’s the case then, the app you are using is the result of style transfer; a computer vision technique that combines your images with artistic style. Introdu...
Neural style transfer and its working
https://nepalprabin.github.io./posts/2020-08-04-general-adversarial-networks-gans.html
General Adversarial Networks (GANs) “General Adversarial Netsis the most interesting idea in the last 10 years in machine learning”. This was the statement from Yann LeCun regarding GANs when Ian Goodfellow and co-authors introduced it in 2014. After its first introduction, many research papers are published with vari...
General Adversarial Networks (GANs)
https://nepalprabin.github.io./posts/2025-03-02-huggingface-smolagents-solutions.html
Huggingface AI Agents Quiz Solutions I have been diving into AI agents through Huggingface’s AI Agents Course. This course offers a comprehensive understanding of how to build and deploy AI agents using thesmolagentslibrary. In this blog, I’ll share insights from the course (Unit 2) and provide code snippets to illust...
Huggingface AI Agents Quiz Solutions
https://nepalprabin.github.io./posts/2021-10-25-autocorrect-and-minimum-edit-distance.html
Autocorrect and Minimum Edit Distance This is my brief note fromDeepLearning.AI’sNLP Specialization Course. What is Autocorrect? Autocorrect is an application that changes misspelled word into a correct word. When writing messages or drafting an email, you may have noticied that if we type any words that is misspell...
Autocorrect and Minimum Edit Distance
https://nepalprabin.github.io./posts/2022-10-19-text-summarization-nlp.html
Text Summarization NLP Text summarization is one of the Natural Language Processing (NLP) tasks where documents/texts are shortened automatically while holding the same semantic meaning. Summarization process generates short, fluent and accurate summary of the long documents. The main idea of text summarization is to ...
Text Summarization NLP
https://nepalprabin.github.io./posts/2021-01-01-deep-residual-learning-for-image-recognition-resnet-paper-explained.html
Deep Residual Learning for Image Recognition (ResNet paper explained) Deep Neural Networks tend to provide more accuracy as the number of layers increases. But, as we go more deeper in the network, the accuracy of the network decreases instead of increasing. As more layers are stacked, there occurs a problem ofvanishi...
Deep Residual Learning for Image Recognition (ResNet paper explained)
https://nepalprabin.github.io./posts/2020-09-21-mobilenet-architecture-explained.html
MobileNet Architecture Explained In this blog post, I will try to write about the MobileNets and its architecture. MobileNet uses depthwise separable convolutions instead of standard convolution to reduce model size and computation. Hence, it can be used to build light weight deep neural networks for mobile and embedd...
MobileNet Architecture Explained
https://nepalprabin.github.io./posts/2020-05-09-vggnet-architecture-explained.html
VGGNet Architecture Explained VGGNet is a Convolutional Neural Network architecture proposed by Karen Simonyan and Andrew Zisserman of University of Oxford in 2014. This paper mailny focuses in the effect of the convolutional neural network depth on its accuracy. You can find the original paper of VGGNet which is titl...
VGGNet Architecture Explained
https://nepalprabin.github.io./posts/2020-04-24-alexnet-architecture-explained.html
AlexNet Architecture Explained AlexNet famously won the 2012 ImageNet LSVRC-2012 competition by a large margin (15.3% vs 26.2%(second place) error rates). Here is the link to originalpaper. Major highlights of the paper - Used ReLU instead of tanh to add non-linearity. - Used dropout instead of regularization to dea...
AlexNet Architecture Explained
https://nepalprabin.github.io./posts/2021-03-26-simclr-explained.html
Paper Explanation: A Simple Framework for Contrastive Learning of Visual Representations (simCLR) Various self-supervised learning methods have been proposed in recent years for learning image representations. Though a lot of methods have been proposed, the performance of those methods was found less effective in term...
Paper Explanation: A Simple Framework for Contrastive Learning of Visual Representations (simCLR)
https://nepalprabin.github.io./posts/2023-07-04-augmented-language-models.html
Augmenting Large Language Models: Expanding Context and Enhancing Relevance With the rise of ChatGPT and other large language models (LLMs), the potential for AI to surpass human capabilities has become a topic of both fascination and concern. While LLMs excel at understanding language, following instructions, and rea...
Augmenting Large Language Models: Expanding Context and Enhancing Relevance
https://nepalprabin.github.io./posts/2020-12-08-self-supervised-learning.html
Self-supervised Learning I have been exploring self-supervised learning and been through papers and blogs to understand it. Self-supervised learning is considered the next big thing in deep learning and why not! If there is a way to learn without providing labels, then this enables us to leverage a large amount of unl...
Self-supervised Learning
https://nepalprabin.github.io./posts/2020-08-15-deep-convolutional-general-adversarial-networks-dcgans.html
Deep Convolutional Generative Adversarial Networks (DCGANs) DCGAN (Deep Convolutional General Adversarial Networks) uses convolutional layers in its design. Architectural Details for DCGAN - Comprised convolutional network without max-pooling. Instead, it uses convolutional stride and transpose convolution for downs...
Deep Convolutional Generative Adversarial Networks (DCGANs)

This contains scraped data from my blog (nepalprabin.github.io)

Downloads last month
7