Home

Awesome

Important Note

Current students of this course at Washington University should refer to the PyTorch version of this course, which is what is currently offered at the university. This repository contains the previous Keras/TensorFlow version.

T81 558:Applications of Deep Neural Networks - TensorFlow

Washington University in St. Louis

Instructor: Jeff Heaton

The content of this course changes as technology evolves, to keep up to date with changes follow me on GitHub.

Course Description

Deep learning is a group of exciting new technologies for neural networks. Through a combination of advanced training techniques and neural network architectural components, it is now possible to create neural networks that can handle tabular data, images, text, and audio as both input and output. Deep learning allows a neural network to learn hierarchies of information in a way that is like the function of the human brain. This course will introduce the student to classic neural network structures, Convolution Neural Networks (CNN), Long Short-Term Memory (LSTM), Gated Recurrent Neural Networks (GRU), General Adversarial Networks (GAN) and reinforcement learning. Application of these architectures to computer vision, time series, security, natural language processing (NLP), and data generation will be covered. High Performance Computing (HPC) aspects will demonstrate how deep learning can be leveraged both on graphical processing units (GPUs), as well as grids. Focus is primarily upon the application of deep learning to problems, with some introduction to mathematical foundations. Students will use the Python programming language to implement deep learning using Google TensorFlow and Keras. It is not necessary to know Python prior to this course; however, familiarity of at least one programming language is assumed. This course will be delivered in a hybrid format that includes both classroom and online instruction.

Textbook

The complete text for this course is here on GitHub. This same material is also available in book format. The course textbook is “Applications of Deep Neural networks with Keras“, ISBN 9798416344269.

If you would like to cite the material from this course/book, please use the following BibTex citation:

@misc{heaton2020applications,
    title={Applications of Deep Neural Networks},
    author={Jeff Heaton},
    year={2020},
    eprint={2009.05673},
    archivePrefix={arXiv},
    primaryClass={cs.LG}
}

Objectives

  1. Explain how neural networks (deep and otherwise) compare to other machine learning models.
  2. Determine when a deep neural network would be a good choice for a particular problem.
  3. Demonstrate your understanding of the material through a final project uploaded to GitHub.

Syllabus

This syllabus presents the expected class schedule, due dates, and reading assignments. Download current syllabus.

ModuleContent
Module 1<br>Meet on 01/23/2023Module 1: Python Preliminaries<ul><li>Part 1.1: Course Overview<li>Part 1.2: Introduction to Python<li>Part 1.3: Python Lists, Dictionaries, Sets & JSON<li>Part 1.4: File Handling<li>Part 1.5: Functions, Lambdas, and Map/ReducePython Preliminaries<li>We will meet on campus this week! (first meeting)</ul>
Module 2<br>Week of 01/30/2023Module 2: Python for Machine Learning<ul><li> Part 2.1: Introduction to Pandas for Deep Learning<li>Part 2.2: Encoding Categorical Values in Pandas<li>Part 2.3: Grouping, Sorting, and Shuffling<li>Part 2.4: Using Apply and Map in Pandas<li>Part 2.5: Feature Engineering in Pandas<li>Module 1 Program due: 01/31/2023<li> Icebreaker due: 01/31/2023</ul>
Module 3<br>Week of 02/06/2023Module 3: TensorFlow and Keras for Neural Networks<ul><li>Part 3.1: Deep Learning and Neural Network Introduction<li>Part 3.2: Introduction to Tensorflow & Keras<li>Part 3.3: Saving and Loading a Keras Neural Network<li>Part 3.4: Early Stopping in Keras to Prevent Overfitting<li>Part 3.5: Extracting Keras Weights and Manual Neural Network Calculation<li>Module 2: Program due: 02/07/2023</ul>
Module 4<br>Week of 02/13/2023Module 4: Training for Tabular Data<ul><li>Part 4.1: Encoding a Feature Vector for Keras Deep Learning<li>Part 4.2: Keras Multiclass Classification for Deep Neural Networks with ROC and AUC<li>Part 4.3: Keras Regression for Deep Neural Networks with RMSE<li>Part 4.4: Backpropagation, Nesterov Momentum, and ADAM Training<li>Part 4.5: Neural Network RMSE and Log Loss Error Calculation from Scratch<li>Module 3 Program due: 02/14/2023</ul>
Module 5<br>Meet on 02/20/2023Module 5: Regularization and Dropout<ul><li>Part 5.1: Introduction to Regularization: Ridge and Lasso<li>Part 5.2: Using K-Fold Cross Validation with Keras<li>Part 5.3: Using L1 and L2 Regularization with Keras to Decrease Overfitting<li>Part 5.4: Drop Out for Keras to Decrease Overfitting<li>Part 5.5: Bootstrapping and Benchmarking Hyperparameters<li>Module 4 Program due: 02/21/2023<li>We will meet on campus this week! (second meeting)</ul>
Module 6<br>Week of 02/27/2023Module 6: CNN for Vision<ul> Part 6.1: Image Processing in Python<li>Part 6.2: Using Convolutional Networks with Keras<li>Part 6.3: Using Pretrained Neural Networks<li>Part 6.4: Looking at Keras Generators and Image Augmentation<li>Part 6.5: Recognizing Multiple Images with YOLOv5<li>Module 5 Program due: 02/28/2023</ul>
Module 7<br>Week of 03/06/2023Module 7: Generative Adversarial Networks (GANs)<ul><li>Part 7.1: Introduction to GANS for Image and Data Generation<li>Part 7.2: Train StyleGAN3 with your Own Images<li>Part 7.3: Exploring the StyleGAN Latent Vector<li>Part 7.4: GANS to Enhance Old Photographs Deoldify<li>Part 7.5: GANs for Tabular Synthetic Data Generation<li>Module 6 Assignment due: 03/07/2023</ul>
Module 8<br>Week of 03/20/2023Module 8: Kaggle<ul><li>Part 8.1: Introduction to Kaggle<li>Part 8.2: Building Ensembles with Scikit-Learn and Keras<li>Part 8.3: How Should you Architect Your Keras Neural Network: Hyperparameters<li>Part 8.4: Bayesian Hyperparameter Optimization for Keras<li>Part 8.5: Current Semester's Kaggle<li>Module 7 Assignment due: 03/21/2023</ul>
Module 9<br>Meet on 03/27/2023Module 9: Transfer Learning<ul><li>Part 9.1: Introduction to Keras Transfer Learning<li>Part 9.2: Keras Transfer Learning for Computer Vision<li>Part 9.3: Transfer Learning for NLP with Keras<li>Part 9.4: Transfer Learning for Facial Feature Recognition<li>Part 9.5: Transfer Learning for Style Transfer<li>We will meet on campus this week! (third meeting)<li>Module 8 Assignment due: 03/28/2023</ul>
Module 10<br>Week of 04/03/2023Module 10: Time Series in Keras<ul><li>Part 10.1: Time Series Data Encoding for Deep Learning, Keras<li>Part 10.2: Programming LSTM with Keras and<li>Part 10.3: Text Generation with Keras<li>Part 10.4: Introduction to Transformers<li>Part 10.5: Transformers for Timeseries<li>Module 9 Assignment due: 04/04/2023</ul>
Module 11<br>Week of 04/10/2023Module 11: Natural Language Processing<ul><li>Part 11.1: Hugging Face Introduction<li>Part 11.2: Hugging Face Tokenizers<li>Part 11.3: Hugging Face Data Sets<li>Part 11.4: Training a Model in Hugging Face<li>Part 11.5: What are Embedding Layers in Keras<li>Module 10 Assignment due: 04/11/2023</ul>
Module 12<br>Week of 04/17/2023Module 12: Reinforcement Learning<ul><li>Kaggle Assignment due: 04/17/2023 (approx 4-6PM, due to Kaggle GMT timezone)<li>Part 12.1: Introduction to the OpenAI Gym<li>Part 12.2: Introduction to Q-Learning for Keras<li>Part 12.3: Keras Q-Learning in the OpenAI Gym<li>Part 12.4: Atari Games with Keras Neural Networks<li>Part 12.5: Application of Reinforcement Learning</ul>
Module 13<br>Meet on 04/24/2023Module 13: Deployment and Monitoring<ul><li>Part 13.1: Flask and Deep Learning Web Services <li>Part 13.2: Interrupting and Continuing Training<li>Part 13.3: Using a Keras Deep Neural Network with a Web Application<li>Part 13.4: When to Retrain Your Neural Network<li>Part 13.5: Tensor Processing Units (TPUs)<li>Final Project due 05/08/2023<li>We will meet on campus this week! (fourth meeting)</ul>

Datasets