Deep Learning for NLP - Part 7

Model Compression for NLP

3.80 (5 reviews)
Udemy
platform
English
language
Other
category
instructor
Deep Learning for NLP - Part 7
82
students
6 hours
content
Aug 2021
last update
$39.99
regular price

What you will learn

Deep Learning for Natural Language Processing

Model Compression for NLP

Pruning

Quantization

Knowledge Distillation

Parameter sharing

Matrix decomposition

DL for NLP

Why take this course?

In recent years, the fields of natural language processing (NLP) and information retrieval (IR) have made tremendous progress thanks to deep learning models like Recurrent Neural Networks (RNNs), Gated Recurrent Units (GRUs) and Long Short-Term Memory (LSTMs) networks, and Transformer based models like Bidirectional Encoder Representations from Transformers (BERT), Generative Pre-training Transformer (GPT-2), Multi-task Deep Neural Network (MT-DNN), Extra-Long Network (XLNet), Text-to-text transfer transformer (T5), T-NLG and GShard.

These models are humongous in size: BERT (340M parameters), GPT-2 (1.5B parameters), T5 (11B parameters, 21.7GB), etc. On the other hand, real world applications demand small model size, low response times and low computational power wattage. In this course, we discuss five different types of methods (Pruning, Quantization, Knowledge Distillation, Parameter Sharing, Tensor Decomposition) for compression of such models to enable their deployment in real industry NLP  projects. Given the critical need of building applications with efficient and small models, and the large amount of recently published work in this area, we believe that this course organizes the plethora of work done by the "deep learning for NLP" community in the past few years and presents it as a coherent story.

Compression for deep learning text models has gained a lot of interest in recent years both from the research community and the industry. Many business owners shy away from using deep learning models fearing the model sizes and infrastructure requirements. Mobile apps need to have a low RAM footprint and clearly a small power envelope. IoT (Internet of Things) and embedded systems related organizations have been investing significantly in designing machine learning solutions for resource constrained environments like sensors.

Researchers in the field of applied deep learning for text will benefit the most, as this tutorial will give them an exhaustive overview of the research in the direction of practical deep learning. We believe that the tutorial will give the newcomers a complete picture of the current work, introduce important research topics in this field, and inspire them to learn more. Practitioners and people from the industry will clearly benefit from the discussions both from the methods perspective, as well from the point of view of applications where such mechanisms are starting to be deployed. This tutorial can be considered an intermediate level tutorial where we assume the folks in audience to know some basic deep learning architectures. Prerequisite knowledge includes introductory level knowledge in deep learning, specifically recurrent neural networks models, and transformers. Also, basic understanding of natural language processing and machine learning concepts is expected.

Screenshots

Deep Learning for NLP - Part 7 - Screenshot_01Deep Learning for NLP - Part 7 - Screenshot_02Deep Learning for NLP - Part 7 - Screenshot_03Deep Learning for NLP - Part 7 - Screenshot_04

Reviews

George
November 12, 2021
Just very high level theory on slides, no practical examples given. No response to the questions or support from the instructor!

Charts

Price

Deep Learning for NLP - Part 7 - Price chart

Rating

Deep Learning for NLP - Part 7 - Ratings chart

Enrollment distribution

Deep Learning for NLP - Part 7 - Distribution chart
4237732
udemy ID
8/12/2021
course created date
8/16/2021
course indexed date
Bot
course submited by