Skip to content

Latest commit

 

History

History
102 lines (64 loc) · 8.07 KB

README.md

File metadata and controls

102 lines (64 loc) · 8.07 KB

AI Curriculum

Open Deep Learning and Reinforcement Learning lectures from top Universities like Stanford University, MIT, UC Berkeley.

Contents

Deep Learning

DS-GA 1008: Deep Learning | SPRING 2020

Lecture Series

This course concerns the latest techniques in deep learning and representation learning, focusing on supervised and unsupervised deep learning, embedding methods, metric learning, convolutional and recurrent nets, with applications to computer vision, natural language understanding, and speech recognition. The lectures are taught by Yann LeCun and the entire lecture series is also available as a YouTube Playlist officially provided by them at https://www.youtube.com/playlist?list=PLLHTzKZzVU9eaEyErdV26ikyolxOsz6mq.

Source: NYU Center for Data Science

Introduction to Deep Learning

MIT 6.S191: Introduction to Deep Learning | 2020

Lecture Series

MIT's introductory course on deep learning methods with applications to computer vision, natural language processing, biology, and more! Students will gain foundational knowledge of deep learning algorithms and get practical experience in building neural networks in TensorFlow. Course concludes with a project proposal competition with feedback from staff and panel of industry sponsors. Prerequisites assume calculus (i.e. taking derivatives) and linear algebra (i.e. matrix multiplication), we'll try to explain everything else along the way! Experience in Python is helpful but not necessary. Listeners are welcome!

Source: MIT

CNNs for Visual Recognition

CS231n: CNNs for Visual Recognition, Stanford | Spring 2019

Lecture Series

Computer Vision has become ubiquitous in our society, with applications in search, image understanding, apps, mapping, medicine, drones, and self-driving cars. Core to many of these applications are visual recognition tasks such as image classification, localization and detection. Recent developments in neural network (aka “deep learning”) approaches have greatly advanced the performance of these state-of-the-art visual recognition systems. This course is a deep dive into details of the deep learning architectures with a focus on learning end-to-end models for these tasks, particularly image classification. During the 10-week course, students will learn to implement, train and debug their own neural networks and gain a detailed understanding of cutting-edge research in computer vision. The final assignment will involve training a multi-million parameter convolutional neural network and applying it on the largest image classification dataset (ImageNet). We will focus on teaching how to set up the problem of image recognition, the learning algorithms (e.g. backpropagation), practical engineering tricks for training and fine-tuning the networks and guide the students through hands-on assignments and a final course project. Much of the background and materials of this course will be drawn from the ImageNet Challenge.

Source: Stanford

NLP with Deep Learning

CS224n: NLP with Deep Learning, Stanford | Winter 2019

Lecture Series

Natural language processing (NLP) is a crucial part of artificial intelligence (AI), modeling how people share information. In recent years, deep learning approaches have obtained very high performance on many NLP tasks. In this course, students gain a thorough introduction to cutting-edge neural networks for NLP.

Source: Stanford

Deep Reinforcement Learning

CS285: Deep Reinforcement Learning, UC Berkeley | Fall 2020

Lecture Series, YouTube

CS285: Deep Reinforcement Learning, UC Berkeley | Fall 2019

Lecture Series, YouTube

Unsupervised Learning

CS294-158-SP20: Deep Unsupervised Learning, UC Berkeley | Spring 2020

Lecture Series, YouTube

This course covers two areas of deep learning in which labeled data is not required: Deep Generative Models and Self-supervised Learning. Recent advances in generative models have made it possible to realistically model high-dimensional raw data such as natural images, audio waveforms and text corpora. Strides in self-supervised learning have started to close the gap between supervised representation learning and unsupervised representation learning in terms of fine-tuning to unseen tasks. This course will cover the theoretical foundations of these topics as well as their newly enabled applications.

Multi-Task and Meta Learning

Stanford CS330: Multi-Task and Meta Learning | 2019

Lecture Series, YouTube

While deep learning has achieved remarkable success in supervised and reinforcement learning problems, such as image classification, speech recognition, and game playing, these models are, to a large degree, specialized for the single task they are trained for. This course will cover the setting where there are multiple tasks to be solved, and study how the structure arising from multiple tasks can be leveraged to learn more efficiently or effectively. This includes:

  • goal-conditioned reinforcement learning techniques that leverage the structure of the provided goal space to learn many tasks significantly faster
  • meta-learning methods that aim to learn efficient learning algorithms that can learn new tasks quickly
  • curriculum and lifelong learning, where the problem requires learning a sequence of tasks, leveraging their shared structure to enable knowledge transfer

Source: Stanford University CS330