ViewTube

ViewTube
Sign inSign upSubscriptions
Filters

Upload date

Type

Duration

Sort by

Features

Reset

3,800 results

CampusX
Layer Normalization in Transformers | Layer Norm Vs Batch Norm

Layer Normalization is a technique used to stabilize and accelerate the training of transformers by normalizing the inputs across ...

46:57
Layer Normalization in Transformers | Layer Norm Vs Batch Norm

61,286 views

1 year ago

Vizuara
Lecture 20: Layer Normalization in the LLM Architecture

In this lecture, we learn about an important component of the LLM architecture: Layer Normalization We understand what exactly ...

38:57
Lecture 20: Layer Normalization in the LLM Architecture

16,651 views

1 year ago

Manning Publications
🧮 Layer Normalization in Transformers – Live Coding with Sebastian Raschka (Chapter 4.2)

Check out Sebastian Raschka's book Build a Large Language Model (From Scratch) | https://hubs.la/Q03l0mSf0 In this ...

22:14
🧮 Layer Normalization in Transformers – Live Coding with Sebastian Raschka (Chapter 4.2)

222 views

6 months ago

Yannic Kilcher
Group Normalization (Paper Explained)

The dirty little secret of Batch Normalization is its intrinsic dependence on the training batch size. Group Normalization attempts to ...

29:06
Group Normalization (Paper Explained)

34,002 views

5 years ago

3Blue1Brown
Attention in transformers, step-by-step | Deep Learning Chapter 6

Demystifying attention, the key mechanism inside transformers and LLMs. Instead of sponsored ad reads, these lessons are ...

26:10
Attention in transformers, step-by-step | Deep Learning Chapter 6

3,619,759 views

1 year ago

Andrej Karpathy
Building makemore Part 3: Activations & Gradients, BatchNorm

We dive into some of the internals of MLPs with multiple layers and scrutinize the statistics of the forward pass activations, ...

1:55:58
Building makemore Part 3: Activations & Gradients, BatchNorm

466,199 views

3 years ago

ExplainingAI
Batch Normalization Explained | Why It Works in Deep Learning

In this video, we dive into Batch Normalization in deep learning, unpacking not just how batch normalization works but also why it ...

38:19
Batch Normalization Explained | Why It Works in Deep Learning

3,628 views

6 months ago

ChiDotPhi
All About Normalizations! - Batch, Layer, Instance and Group Norm

In this video, I review the different kinds of normalizations used in Deep Learning. Note, I accidentally interchange std and ...

41:56
All About Normalizations! - Batch, Layer, Instance and Group Norm

19,287 views

3 years ago

NPTEL IIT Delhi
Lec 16 | Introduction to Transformer: Positional Encoding and Layer Normalization

This lecture dives into the technical aspects of positional encoding methods and layer normalization within the Transformer ...

1:26:53
Lec 16 | Introduction to Transformer: Positional Encoding and Layer Normalization

7,151 views

11 months ago

Michigan Online
Lecture 7: Convolutional Networks

Lecture 7 moves from fully-connected to convolutional networks by introducing new computational primitives that respect the ...

1:08:53
Lecture 7: Convolutional Networks

74,870 views

5 years ago

Yannic Kilcher
Evolving Normalization-Activation Layers

Normalization and activation layers have seen a long history of hand-crafted variants with various results. This paper proposes an ...

30:38
Evolving Normalization-Activation Layers

2,914 views

5 years ago

Unfold Data Science
Residual Connections and Layer Normalization |Layer Normalization vs Batch Normalization|Transformer

Residual Connections and Layer Normalization |Layer Normalization vs Batch Normalization |Tranformer #ai #artificialintelligence ...

40:37
Residual Connections and Layer Normalization |Layer Normalization vs Batch Normalization|Transformer

1,021 views

8 months ago

LCS2
LLMs | Intro to Transformer: Positional Encoding and Layer Normalization  | Lec 6.2

tl;dr: This lecture dives into the technical aspects of positional encoding methods and layer normalization within the Transformer ...

1:27:44
LLMs | Intro to Transformer: Positional Encoding and Layer Normalization | Lec 6.2

5,721 views

1 year ago

MLWorks
🚀 Cuda Programming Day 5: Layer Normalization | Neural Network | Transformer Architecture

Welcome to CUDA Programming Day 5! Today, we step into the world of deep learning and explore how CUDA powers one of the ...

23:16
🚀 Cuda Programming Day 5: Layer Normalization | Neural Network | Transformer Architecture

105 views

1 month ago

McData
Transformers Explained: Layer Normalization and Residual Connections

This video explains the basics of layer normalization and residual connection---both are building blocks of a transformer model.

22:32
Transformers Explained: Layer Normalization and Residual Connections

12 views

2 weeks ago

Santanu Kundu
Lecture 49  Layer, Instance, Group Normalization

A Deep Learning Discussion by Dr. Prabir Kumar Biswas, A renowned professor of Electronics and Electrical Communication ...

28:43
Lecture 49 Layer, Instance, Group Normalization

4,380 views

6 years ago

Microsoft Research
How does Batch Normalization Help Optimization?

Batch normalization (BatchNorm) is a widely adopted technique that enables faster and more stable training of deep neural ...

48:05
How does Batch Normalization Help Optimization?

10,228 views

6 years ago

Yannic Kilcher
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift

https://arxiv.org/abs/1502.03167 Abstract: Training Deep Neural Networks is complicated by the fact that the distribution of each ...

25:44
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift

27,817 views

6 years ago

sudoremove
Understanding Normalization [Batch Norm, Layer Norm, Instance Norm]

Chapter --- 00:00 Intro 03:17 Normalization is Appearing So Much 01:46 The Emergence of BatchNorm: The Era of CNN ...

27:27
Understanding Normalization [Batch Norm, Layer Norm, Instance Norm]

1,183 views

2 weeks ago

Data Science Courses
Ali Ghodsi, Deep Learning,  Regularization (Layer norm, FRN,TRU), Keras, Fall 2023, Lecture 7

Layer normalization, Filter response normalization (FRN), Thresholded linear unit (TLU), Normalizer-free networks, Gradient ...

52:03
Ali Ghodsi, Deep Learning, Regularization (Layer norm, FRN,TRU), Keras, Fall 2023, Lecture 7

1,984 views

2 years ago