ViewTube

ViewTube
Sign inSign upSubscriptions
Filters

Upload date

Type

Duration

Sort by

Features

Reset

184 results

CampusX
Layer Normalization in Transformers | Layer Norm Vs Batch Norm

Layer Normalization is a technique used to stabilize and accelerate the training of transformers by normalizing the inputs across ...

46:57
Layer Normalization in Transformers | Layer Norm Vs Batch Norm

60,602 views

1 year ago

Data Science Courses
Ali Ghodsi, Deep Learning,  Regularization (Layer norm, FRN,TRU), Keras, Fall 2023, Lecture 7

Layer normalization, Filter response normalization (FRN), Thresholded linear unit (TLU), Normalizer-free networks, Gradient ...

52:03
Ali Ghodsi, Deep Learning, Regularization (Layer norm, FRN,TRU), Keras, Fall 2023, Lecture 7

1,979 views

2 years ago

Jeremy Howard
Lesson 17: Deep Learning Foundations to Stable Diffusion

We explore normalization techniques, such as Layer Normalization and Batch Normalization, and briefly mention other ...

1:56:33
Lesson 17: Deep Learning Foundations to Stable Diffusion

11,869 views

2 years ago

CampusX
Batch Normalization in Deep Learning | Batch Learning in Keras

This video explores how Batch Normalization transforms the internal workings of neural networks by normalizing inputs within ...

43:39
Batch Normalization in Deep Learning | Batch Learning in Keras

108,705 views

3 years ago

CodeEmporium
[ 100k Special ] Transformers: Zero to Hero

0:29 Transformer Overview 12:27 Self Attention 26:40 Multihead Attention 39:31 Position Encoding 48:51 Layer Normalization ...

3:34:41
[ 100k Special ] Transformers: Zero to Hero

67,655 views

2 years ago

Derek Harter
L11.4.3-4: Transformer Architecture: Implementing and adding positional encoding

It includes several Dense layers to factor outputs into multiple independent spaces. Also concepts like layer normalization and ...

26:52
L11.4.3-4: Transformer Architecture: Implementing and adding positional encoding

21 views

6 months ago

Data Science Courses
Ali Ghodsi, Deep Learning,  Dropout, Batch Normalization, Fall 2023, Lecture 5

Dropout, Batch normalization Batch normalization was initially inspired by the notion of internal covariate shift (ICS). However, it's ...

1:07:27
Ali Ghodsi, Deep Learning, Dropout, Batch Normalization, Fall 2023, Lecture 5

3,159 views

2 years ago

Derek Harter
L09.3: Modern Convnet Architecture Patterns

In this video I will discuss some of the more advanced features and best practices that you will find in more recent and modern ...

27:32
L09.3: Modern Convnet Architecture Patterns

22 views

7 months ago

CodeEmporium
Blowing up the Transformer Encoder!

10:53 Combining Attention heads 12:46 Residual Connections (Skip Connections) 13:45 Layer Normalization 16:36 Why Linear ...

20:58
Blowing up the Transformer Encoder!

24,398 views

2 years ago

moccam1
Toolkait - presentation of the interface

This video presents the various Toolkait modules that allow you to load data (numbers, categories, images, text, etc.), process it ...

51:08
Toolkait - presentation of the interface

6 views

1 month ago

Neural Reckoning
Iulia M. Comsa (Google Research) - On temporal coding in SNNs with alpha synaptic function

The timing of individual neuronal spikes is essential for biological brains to make fast responses to sensory stimuli. However ...

52:19
Iulia M. Comsa (Google Research) - On temporal coding in SNNs with alpha synaptic function

1,958 views

5 years ago

Wuttipong วุฒิพงษ์ Kumwilaisak คําวิลัยศักดิ์
Deep Learning Module 2 Part 3/1: Normalizing Inputs
21:37
Deep Learning Module 2 Part 3/1: Normalizing Inputs

219 views

5 years ago

Alfredo Canziani (冷在)
Week 14 – Practicum: Overfitting and regularization, and Bayesian neural nets

Course website: http://bit.ly/pDL-home Playlist: http://bit.ly/pDL-YouTube Speaker: Alfredo Canziani Week 14: ...

1:11:28
Week 14 – Practicum: Overfitting and regularization, and Bayesian neural nets

5,295 views

5 years ago

Data Science Learning Community Videos
Practical Deep Learning for Coders: Initialization/normalization (pdl01 17)

Aaron G leads a discussion of Chapter 17 ("Initialization/normalization") from Practical Deep Learning for Coders by Jeremy ...

52:16
Practical Deep Learning for Coders: Initialization/normalization (pdl01 17)

65 views

1 year ago

Alfredo Canziani (冷在)
13L – Optimisation for Deep Learning

Course website: http://bit.ly/DLSP21-web Playlist: http://bit.ly/DLSP21-YouTube Speaker: Yann LeCun Chapters 00:00:00 ...

1:51:33
13L – Optimisation for Deep Learning

7,694 views

4 years ago

Derek Harter
L11.3.2: Two Approaches for Representing Groups of Words: Bag-of-Word Models

In this video I look at the two basic approaches that you can use to represent word order for processing text with a deep network ...

23:38
L11.3.2: Two Approaches for Representing Groups of Words: Bag-of-Word Models

12 views

6 months ago

CodeEmporium
Transformer Decoder coded from scratch

... 9:07 Decoder Forward Pass 11:28 Decoder Layer 13:00 Masked Multi Head Self Attention 23:00 Dropout + Layer Normalization ...

39:54
Transformer Decoder coded from scratch

13,656 views

2 years ago

Alfredo Canziani (冷在)
Week 12 – Practicum: Attention and the Transformer

Course website: http://bit.ly/DLSP20-web Playlist: http://bit.ly/pDL-YouTube Speaker: Alfredo Canziani Week 12: ...

1:18:02
Week 12 – Practicum: Attention and the Transformer

13,049 views

5 years ago

Minjoon Seo
Deep Learning for NLP - Lec 07 (KAIST AI605 Spring 2022)

Course website: https://seominjoon.github.io/kaist-ai605/

54:10
Deep Learning for NLP - Lec 07 (KAIST AI605 Spring 2022)

180 views

3 years ago

CodeEmporium
Transformer Encoder in 100 lines of code!

... OF VIDEO: “MultiHeadAttention” Class 36:27 Returning the flow back to “EncoderLayer” Class 37:12 Layer Normalization 43:17 ...

49:54
Transformer Encoder in 100 lines of code!

22,086 views

2 years ago