Classbaze

Disclosure: when you buy through links on our site, we may earn an affiliate commission.

Deep Learning for NLP – Part 6

Part 6: Popular Transformer Models
0.0
0.0/5
(0 reviews)
41 students
Created by

8.6

Classbaze Grade®

9.2

Freshness

N/A

Popularity

7.6

Material

Part 6: Popular Transformer Models
Platform: Udemy
Video: 2h 39m
Language: English
Next start: On Demand

Best Deep Learning classes:

Classbaze Rating

Classbaze Grade®

8.6 / 10

CourseMarks Score® helps students to find the best classes. We aggregate 18 factors, including freshness, student feedback and content diversity.

Freshness

9.2 / 10
This course was last updated on 8/2021.

Course content can become outdated quite quickly. After analysing 71,530 courses, we found that the highest rated courses are updated every year. If a course has not been updated for more than 2 years, you should carefully evaluate the course before enrolling.

Popularity

We analyzed factors such as the rating (0.0/5) and the ratio between the number of reviews and the number of students, which is a great signal of student commitment.

New courses are hard to evaluate because there are no or just a few student ratings, but Student Feedback Score helps you find great courses even with fewer reviews.

Material

7.6 / 10
Video Score: 8.0 / 10
The course includes 2h 39m video content. Courses with more videos usually have a higher average rating. We have found that the sweet spot is 16 hours of video, which is long enough to teach a topic comprehensively, but not overwhelming. Courses over 16 hours of video gets the maximum score.
The average video length is 8 hours 18 minutes of 153 Deep Learning courses on Udemy.
Detail Score: 9.2 / 10

The top online course contains a detailed description of the course, what you will learn and also a detailed description about the instructor.

Extra Content Score: 5.5 / 10

Tests, exercises, articles and other resources help students to better understand and deepen their understanding of the topic.

This course contains:

0 article.
0 resource.
0 exercise.
0 test.

In this page

About the course

This course is a part of “Deep Learning for NLP” Series. In this course, I will talk about various popular Transformer models beyond the ones I have already covered in the previous sessions in this series. Such Transformer models including encoder as well as decoder based models and differ in terms of various aspects like form of input, pretraining objectives, pretraining data, architecture variations, etc.
These Transformer models have been all proposed after 2019 and some of them are also from early 2021. Thus, as of Aug 2021, these models are very recent and state of the art across multiple NLP tasks.
The course consists of three main sections as follows.
In the first section, I will talk about a few Transformer encoder and decoder models which extend the original Transformer framework. Specifically I will cover SpanBERT, Electra, DeBERTa and DialoGPT. SpanBERT, Electra and DeBERTa are Transformer encoders while DialoGPT is a Transformer decoder model. For each model, we will also talk about their architecture or pretraining differs from standard Transformer. We will also talk important results on various NLP tasks.
In the second section, I will talk about multi-modal Transformer models. Multimodal learning has gained a lot of momentum in recent years. Thus, there was a need to come up with Transformer models which could handle text and image data together. In this part, I will cover VisualBERT and vilBERT which both process the multi-modal input very effectively. Both the models have many similarities. We will discuss about theri similarities and differences in detail.
Lastly, in the third section, I will talk about lareg scale Transformer models. I will introduce the mixture of experts (MoE) architecture. Then I will talk about how  GShard adapts the MoE architecture, and shows great results on massive multilingual machine translation. Lastly, I will discuss Switch Transformers which simplify the MoE routing algorithm and also do several engineering optimizations to reduce network communciation and computation costs and mitigate instabilities.
In general, each of these papers is pretty long and thus it becomes very difficult and time consuming to understand them. In these sessions, I have tried to summarize them nicely bringing out the intuitions and tying the important concepts across such papers in a coherent story. Hope you will find it useful for your work and understanding.

What can you learn from this course?

✓ Deep Learning for Natural Language Processing
✓ Popular Transformer encoder and decoder models
✓ Multi-modal Transformer models
✓ Large scale Transformer models
✓ DL for NLP

What you need to start the course?

• Basics of machine learning
• Basic understanding of Transformer based models and word embeddings
• Transformer Models like BERT and GPT

Who is this course is made for?

• Beginners in deep learning
• Python developers interested in data science concepts
• Masters or PhD students who wish to learn deep learning concepts quickly

Are there coupons or discounts for Deep Learning for NLP - Part 6 ? What is the current price?

The course costs $14.99. And currently there is a 25% discount on the original price of the course, which was $19.99. So you save $5 if you enroll the course now.
The average price is $16.2 of 153 Deep Learning courses. So this course is 7% cheaper than the average Deep Learning course on Udemy.

Will I be refunded if I'm not satisfied with the Deep Learning for NLP - Part 6 course?

YES, Deep Learning for NLP – Part 6 has a 30-day money back guarantee. The 30-day refund policy is designed to allow students to study without risk.

Are there any financial aid for this course?

Currently we could not find a scholarship for the Deep Learning for NLP - Part 6 course, but there is a $5 discount from the original price ($19.99). So the current price is just $14.99.

Who will teach this course? Can I trust Manish Gupta?

Manish Gupta has created 10 courses that got 29 reviews which are generally positive. Manish Gupta has taught 177 students and received a 4.3 average review out of 29 reviews. Depending on the information available, we think that Manish Gupta is an instructor that you can trust.
Principal Applied Researcher
Manish Gupta is a Principal Applied Researcher at Microsoft India R&D Private Limited at Hyderabad, India. He is also an Adjunct Faculty at International Institute of Information Technology, Hyderabad and a visiting faculty at Indian School of Business, Hyderabad. He received his Masters in Computer Science from IIT Bombay in 2007 and his Ph.D. from the University of Illinois at Urbana-Champaign in 2013. Before this, he worked for Yahoo! Bangalore for two years. His research interests are in the areas of web mining, data mining and information retrieval. He has published more than 100 research papers in reputed refereed journals and conferences. He has also co-authored two books: one on Outlier Detection for Temporal Data and another one on Information Retrieval with Verbose Queries.

8.6

Classbaze Grade®

9.2

Freshness

N/A

Popularity

7.6

Material

Platform: Udemy
Video: 2h 39m
Language: English
Next start: On Demand

Classbaze recommendations for you