Category: Deep Learning Tutorials


  • What Is Dropout Layer?

    What Is Dropout Layer?

    What Is Dropout Layer? Table Of Contents: What Is Dropout Layer? What Happens In Training Stage ? What Happens In Testing Stage ? Why We Need To Scale The Weights After Training, When Using Dropouts? (1) What Is Dropout Layer ? The Dropout Layer is a regularization technique used in deep learning neural networks to prevent overfitting. Overfitting occurs when a model performs exceptionally well on the training data but fails to generalize well to new, unseen data. The Dropout Layer works by randomly “dropping out” (i.e., temporarily deactivating) a proportion of the neurons in a neural network during the

    Read More

  • What Is Early Stopping?

    What Is Early Stopping?

    What Is Early Stopping ? Table Of Contents: What Is Early Stopping? Example To Understand – Classification Use Case. Understand The EarlyStopping() Method. (1) What Is Early Stopping? Let’s say you are training a neural network model, you need to mention how many epochs you need to train your model. the term “epochs” refers to a single complete pass of the training dataset through the neural network. How would you know how many epochs you need to have to train your model perfectly? You can say that I will train my model 1000 thousand times and see the result. But

    Read More

  • Self Attentions In Transformers.

    Self Attentions In Transformers.

    Self Attention In Transformers Table Of Contents: Motivation To Study Self Attention. Problem With Word Embedding. What Is Contextual Word Embedding? How Does Self-Attention Work? How To Get The Contextual Word Embeddings? Advantages Of First Principle Above Approach. Introducing Learnable Parameters In The Model. (1) Motivation To Study Self Attention. In 2024 we all know that there is a technology called ‘GenAI’ has penetrated into the market. With this technology we can create different new images, videos, texts from scratch automatically. The center of ‘GenAI’ technology is the ‘Transformers’. And the center of the Transformer is the ‘Self Attention’. Hence

    Read More

  • What Is Self Attention ?

    What Is Self Attention ?

    What Is Self Attention ? Table Of Contents: What Is The Most Important Thing In NLP Applications? Problem With Word2Vec Model. The Problem Of Average Meaning. What Is Self Attention? (1) What Is The Most Important Thing In NLP Applications? Before understanding the self-attention mechanism, we must understand the most important thing in any NLP application. The answer is how you convert any words into numbers ? Our computers don’t understand words they only understand numbers. Hence the researchers first worked in this direction to convert any words into vectors. We got some basic techniques like, One Hot Encoding. Bag

    Read More

  • Introduction To Transformers!

    Introduction To Transformers!

    Introduction To Transformers ! Table Of Contents: What Is Transformers? History Of Transformers. Impact Of Transformers In NLP. Democratizing AI. Multimodel Capability Of Transformers. Acceleration Of GenAI. Unification Of Deep Learning. Why Transformers Are Created? Neural Machine Translation Jointly Learning To Align & Translate. Attention Is All You Need. The Time Line Of Transformers. The Advantages Of Transformers. Real World Applications Of Transformers. Disadvantages Of Transformers. The Future Of Transformers. (1) What Is Transformers? Transformers is basically a Neural Network Architecture. In deep learning, we have already studied the ANN, CNN & RNN. ANN works for the cross-sectional data, CNN

    Read More

  • Luong Attention !

    Luong Attention !

    Luong’s Attention ! Table Of Contents: What Is Luong’s Attention?  Key Features Of Luong’s Attention Model? Advantages Of Luong’s Attention Model? Architecture Of Luong’s Attention Model.  Why do We Take the Current Hidden State Output Of The Decoder In Luong’s Attention Model?  Architecture Luong’s Attention Model. Difference In Luong’s Attention & Bahdanau’s Attention (1) What Is Luong’s Attention? Luong’s attention is another type of attention mechanism, introduced in the paper “Effective Approaches to Attention-based Neural Machine Translation” by Minh-Thang Luong, Hieu Pham, and Christopher D. Manning in 2015. Luong’s attention mechanism is also designed for encoder-decoder models, similar to Bahdanau’s

    Read More

  • Bahdanau Attention Vs Luong Attention !

    Bahdanau Attention Vs Luong Attention !

    Bahdanau Attention ! Table Of Contents: What Is Attention Mechanism? What Is Bahdanau’s Attention? Architecture Of Bahdanau’s Attention? (1) What Is Attention Mechanism? An attention mechanism is a neural network component used in various deep learning models, particularly in the field of natural language processing (NLP) and sequence-to-sequence tasks. It was introduced in the paper “Attention is All You Need” by Vaswani et al. in 2017. The attention mechanism allows a model to focus on the most relevant parts of the input when generating an output, rather than treating the entire input sequence equally. This is particularly useful when the

    Read More

  • What Is Attention Mechanism?

    What Is Attention Mechanism?

    What Is Attention Mechanism? Table Of Contents: Problem With Encoder & Decoder Architecture. Solution For Encoder & Decoder Architecture. Math’s Behind Attention Mechanism. Improvements Due To Attention Mechanism. (1) Problem With Encoder & Decoder Architecture. Problem With Encoder: The main idea behind the Encoder & Decoder architecture is that, the encoder summarize the entire text into one vector format and from that vector we need to convert into different language. Let us consider the below example, Your task is to read the entire sentence first, keep all the words in mind and translate it into Hindi without seeing the sentence

    Read More

  • Encoder Decoder Architecture

    Encoder Decoder Architecture

    Encoder Decoder Architecture Table Of Contents: Milestones In Deep Learning. What Is Sequence To Sequence Model? Problem With LSTM Networks. Challenges To Handle Sequence To Sequence Data. How Does LSTM Network Can Handle Variable Length Input ? How Does LSTM Network Can Handle Variable Length Output With Some Trick? Why Does The LSTM Don’t Have The Decision Making Capability ? Challenges To Handle Sequence To Sequence Data. High-level Overview Of Encoder Decoder Architecture. What Is Inside Encoder Module? What Is Inside Decoder Module? How To Train Encoder & Decoder Architecture. Model Prediction (1) Milestones In Deep Learning ? Milestone-1: In

    Read More

  • Word2Vec Word Vectorization

    Word2Vec Word Vectorization

    Word2Vec Word Vectorization Table Of Contents: What Is Word Embedding? Types Of Word Embedding. What Is Word2Vec? Why Are Word Embedding Needed? How Word2Vec Model Works? Pretrained Word2Vec Model. What Are 300 Dimension Numbers Signifies? Intuition Behind Word2Vec. Assumption Behind Word2Vec Mode. Architecture Of Word2Vec Model. Continuous Bag Of Words(CBOW). Skip-Gram Word2Vec. When To Use CBOW & Skip-Gram? How To Increase The Performance Of The Word2Vec Model? Train Word2Vec Model With Game Of Thrones Dataset. (1) What Is Word Embedding? Word embedding is a fundamental technique in natural language processing (NLP) that represents words as dense, low-dimensional vectors of real

    Read More