PyTorch for Deep Learning
52h 27s
English
Paid
Course description
Learn PyTorch from scratch! This PyTorch course is your step-by-step guide to developing your own deep learning models using PyTorch. You'll learn Deep Learning with PyTorch by building a massive 3-part real-world milestone project. By the end, you'll have the skills and portfolio to get hired as a Deep Learning Engineer.
Watch Online
0:00
/ #1: PyTorch for Deep Learning
All Course Lessons (348)
| # | Lesson Title | Duration | Access |
|---|---|---|---|
| 1 | PyTorch for Deep Learning Demo | 03:34 | |
| 2 | Course Welcome and What Is Deep Learning | 05:54 | |
| 3 | Why Use Machine Learning or Deep Learning | 03:34 | |
| 4 | The Number 1 Rule of Machine Learning and What Is Deep Learning Good For | 05:40 | |
| 5 | Machine Learning vs. Deep Learning | 06:07 | |
| 6 | Anatomy of Neural Networks | 09:22 | |
| 7 | Different Types of Learning Paradigms | 04:31 | |
| 8 | What Can Deep Learning Be Used For | 06:22 | |
| 9 | What Is and Why PyTorch | 10:13 | |
| 10 | What Are Tensors | 04:16 | |
| 11 | What We Are Going To Cover With PyTorch | 06:06 | |
| 12 | How To and How Not To Approach This Course | 05:10 | |
| 13 | Important Resources For This Course | 05:22 | |
| 14 | Getting Setup to Write PyTorch Code | 07:40 | |
| 15 | Introduction to PyTorch Tensors | 13:26 | |
| 16 | Creating Random Tensors in PyTorch | 09:59 | |
| 17 | Creating Tensors With Zeros and Ones in PyTorch | 03:09 | |
| 18 | Creating a Tensor Range and Tensors Like Other Tensors | 05:18 | |
| 19 | Dealing With Tensor Data Types | 09:25 | |
| 20 | Getting Tensor Attributes | 08:23 | |
| 21 | Manipulating Tensors (Tensor Operations) | 06:00 | |
| 22 | Matrix Multiplication (Part 1) | 09:35 | |
| 23 | Matrix Multiplication (Part 2): The Two Main Rules of Matrix Multiplication | 07:52 | |
| 24 | Matrix Multiplication (Part 3): Dealing With Tensor Shape Errors | 12:58 | |
| 25 | Finding the Min Max Mean and Sum of Tensors (Tensor Aggregation) | 06:10 | |
| 26 | Finding The Positional Min and Max of Tensors | 03:17 | |
| 27 | Reshaping, Viewing and Stacking Tensors | 13:41 | |
| 28 | Squeezing, Unsqueezing and Permuting Tensors | 11:56 | |
| 29 | Selecting Data From Tensors (Indexing) | 09:32 | |
| 30 | PyTorch Tensors and NumPy | 09:09 | |
| 31 | PyTorch Reproducibility (Taking the Random Out of Random) | 10:47 | |
| 32 | Different Ways of Accessing a GPU in PyTorch | 11:51 | |
| 33 | Setting up Device Agnostic Code and Putting Tensors On and Off the GPU | 07:44 | |
| 34 | PyTorch Fundamentals: Exercises and Extra-Curriculum | 04:50 | |
| 35 | Introduction and Where You Can Get Help | 02:46 | |
| 36 | Getting Setup and What We Are Covering | 07:15 | |
| 37 | Creating a Simple Dataset Using the Linear Regression Formula | 09:42 | |
| 38 | Splitting Our Data Into Training and Test Sets | 08:21 | |
| 39 | Building a function to Visualize Our Data | 07:46 | |
| 40 | Creating Our First PyTorch Model for Linear Regression | 14:10 | |
| 41 | Breaking Down What's Happening in Our PyTorch Linear regression Model | 06:11 | |
| 42 | Discussing Some of the Most Important PyTorch Model Building Classes | 06:27 | |
| 43 | Checking Out the Internals of Our PyTorch Model | 09:51 | |
| 44 | Making Predictions With Our Random Model Using Inference Mode | 11:13 | |
| 45 | Training a Model Intuition (The Things We Need) | 08:15 | |
| 46 | Setting Up an Optimizer and a Loss Function | 12:52 | |
| 47 | PyTorch Training Loop Steps and Intuition | 13:54 | |
| 48 | Writing Code for a PyTorch Training Loop | 08:47 | |
| 49 | Reviewing the Steps in a Training Loop Step by Step | 14:58 | |
| 50 | Running Our Training Loop Epoch by Epoch and Seeing What Happens | 09:26 | |
| 51 | Writing Testing Loop Code and Discussing What's Happening Step by Step | 11:38 | |
| 52 | Reviewing What Happens in a Testing Loop Step by Step | 14:43 | |
| 53 | Writing Code to Save a PyTorch Model | 13:46 | |
| 54 | Writing Code to Load a PyTorch Model | 08:45 | |
| 55 | Setting Up to Practice Everything We Have Done Using Device-Agnostic Code | 06:03 | |
| 56 | Putting Everything Together (Part 1): Data | 06:09 | |
| 57 | Putting Everything Together (Part 2): Building a Model | 10:08 | |
| 58 | Putting Everything Together (Part 3): Training a Model | 12:41 | |
| 59 | Putting Everything Together (Part 4): Making Predictions With a Trained Model | 05:18 | |
| 60 | Putting Everything Together (Part 5): Saving and Loading a Trained Model | 09:11 | |
| 61 | Exercise: Imposter Syndrome | 02:57 | |
| 62 | PyTorch Workflow: Exercises and Extra-Curriculum | 03:58 | |
| 63 | Introduction to Machine Learning Classification With PyTorch | 09:42 | |
| 64 | Classification Problem Example: Input and Output Shapes | 09:08 | |
| 65 | Typical Architecture of a Classification Neural Network (Overview) | 06:32 | |
| 66 | Making a Toy Classification Dataset | 12:19 | |
| 67 | Turning Our Data into Tensors and Making a Training and Test Split | 11:56 | |
| 68 | Laying Out Steps for Modelling and Setting Up Device-Agnostic Code | 04:20 | |
| 69 | Coding a Small Neural Network to Handle Our Classification Data | 10:58 | |
| 70 | Making Our Neural Network Visual | 06:58 | |
| 71 | Recreating and Exploring the Insides of Our Model Using nn.Sequential | 13:18 | |
| 72 | Setting Up a Loss Function Optimizer and Evaluation Function for Our Classification Network | 14:51 | |
| 73 | Going from Model Logits to Prediction Probabilities to Prediction Labels | 16:08 | |
| 74 | Coding a Training and Testing Optimization Loop for Our Classification Model | 15:27 | |
| 75 | Writing Code to Download a Helper Function to Visualize Our Models Predictions | 14:14 | |
| 76 | Discussing Options to Improve a Model | 08:03 | |
| 77 | Creating a New Model with More Layers and Hidden Units | 09:07 | |
| 78 | Writing Training and Testing Code to See if Our New and Upgraded Model Performs Better | 12:46 | |
| 79 | Creating a Straight Line Dataset to See if Our Model is Learning Anything | 08:08 | |
| 80 | Building and Training a Model to Fit on Straight Line Data | 10:02 | |
| 81 | Evaluating Our Models Predictions on Straight Line Data | 05:24 | |
| 82 | Introducing the Missing Piece for Our Classification Model Non-Linearity | 10:01 | |
| 83 | Building Our First Neural Network with Non-Linearity | 10:26 | |
| 84 | Writing Training and Testing Code for Our First Non-Linear Model | 15:13 | |
| 85 | Making Predictions with and Evaluating Our First Non-Linear Model | 05:48 | |
| 86 | Replicating Non-Linear Activation Functions with Pure PyTorch | 09:35 | |
| 87 | Putting It All Together (Part 1): Building a Multiclass Dataset | 11:25 | |
| 88 | Creating a Multi-Class Classification Model with PyTorch | 12:28 | |
| 89 | Setting Up a Loss Function and Optimizer for Our Multi-Class Model | 06:41 | |
| 90 | Going from Logits to Prediction Probabilities to Prediction Labels with a Multi-Class Model | 11:03 | |
| 91 | Training a Multi-Class Classification Model and Troubleshooting Code on the Fly | 16:18 | |
| 92 | Making Predictions with and Evaluating Our Multi-Class Classification Model | 08:00 | |
| 93 | Discussing a Few More Classification Metrics | 09:18 | |
| 94 | PyTorch Classification: Exercises and Extra-Curriculum | 02:59 | |
| 95 | What Is a Computer Vision Problem and What We Are Going to Cover | 11:48 | |
| 96 | Computer Vision Input and Output Shapes | 10:09 | |
| 97 | What Is a Convolutional Neural Network (CNN) | 05:03 | |
| 98 | Discussing and Importing the Base Computer Vision Libraries in PyTorch | 09:20 | |
| 99 | Getting a Computer Vision Dataset and Checking Out Its- Input and Output Shapes | 14:31 | |
| 100 | Visualizing Random Samples of Data | 09:52 | |
| 101 | DataLoader Overview Understanding Mini-Batch | 07:18 | |
| 102 | Turning Our Datasets Into DataLoaders | 12:24 | |
| 103 | Model 0: Creating a Baseline Model with Two Linear Layers | 14:39 | |
| 104 | Creating a Loss Function: an Optimizer for Model 0 | 10:30 | |
| 105 | Creating a Function to Time Our Modelling Code | 05:35 | |
| 106 | Writing Training and Testing Loops for Our Batched Data | 21:26 | |
| 107 | Writing an Evaluation Function to Get Our Models Results | 12:59 | |
| 108 | Setup Device-Agnostic Code for Running Experiments on the GPU | 03:47 | |
| 109 | Model 1: Creating a Model with Non-Linear Functions | 09:04 | |
| 110 | Mode 1: Creating a Loss Function and Optimizer | 03:05 | |
| 111 | Turing Our Training Loop into a Function | 08:29 | |
| 112 | Turing Our Testing Loop into a Function | 06:36 | |
| 113 | Training and Testing Model 1 with Our Training and Testing Functions | 11:53 | |
| 114 | Getting a Results Dictionary for Model 1 | 04:09 | |
| 115 | Model 2: Convolutional Neural Networks High Level Overview | 08:25 | |
| 116 | Model 2: Coding Our First Convolutional Neural Network with PyTorch | 19:49 | |
| 117 | Model 2: Breaking Down Conv2D Step by Step | 15:00 | |
| 118 | Model 2: Breaking Down MaxPool2D Step by Step | 15:49 | |
| 119 | Mode 2: Using a Trick to Find the Input and Output Shapes of Each of Our Layers | 13:46 | |
| 120 | Model 2: Setting Up a Loss Function and Optimizer | 02:39 | |
| 121 | Model 2: Training Our First CNN and Evaluating Its Results | 07:55 | |
| 122 | Comparing the Results of Our Modelling Experiments | 07:24 | |
| 123 | Making Predictions on Random Test Samples with the Best Trained Model | 11:40 | |
| 124 | Plotting Our Best Model Predictions on Random Test Samples and Evaluating Them | 08:11 | |
| 125 | Making Predictions Across the Whole Test Dataset and Importing Libraries to Plot a Confusion Matrix | 15:21 | |
| 126 | Evaluating Our Best Models Predictions with a Confusion Matrix | 06:55 | |
| 127 | Saving and Loading Our Best Performing Model | 11:28 | |
| 128 | Recapping What We Have Covered Plus Exercises and Extra-Curriculum | 06:02 | |
| 129 | What Is a Custom Dataset and What We Are Going to Cover | 09:54 | |
| 130 | Importing PyTorch and Setting Up Device-Agnostic Code | 05:55 | |
| 131 | Downloading a Custom Dataset of Pizza, Steak and Sushi Images | 14:05 | |
| 132 | Becoming One With the Data (Part 1): Exploring the Data Format | 08:42 | |
| 133 | Becoming One With the Data (Part 2): Visualizing a Random Image | 11:41 | |
| 134 | Becoming One With the Data (Part 3): Visualizing a Random Image with Matplotlib | 04:48 | |
| 135 | Transforming Data (Part 1): Turning Images Into Tensors | 08:54 | |
| 136 | Transforming Data (Part 2): Visualizing Transformed Images | 11:31 | |
| 137 | Loading All of Our Images and Turning Them Into Tensors With ImageFolder | 09:19 | |
| 138 | Visualizing a Loaded Image From the Train Dataset | 07:19 | |
| 139 | Turning Our Image Datasets into PyTorch DataLoaders | 09:04 | |
| 140 | Creating a Custom Dataset Class in PyTorch High Level Overview | 08:01 | |
| 141 | Creating a Helper Function to Get Class Names From a Directory | 09:07 | |
| 142 | Writing a PyTorch Custom Dataset Class from Scratch to Load Our Images | 17:47 | |
| 143 | Compare Our Custom Dataset Class to the Original ImageFolder Class | 07:14 | |
| 144 | Writing a Helper Function to Visualize Random Images from Our Custom Dataset | 14:19 | |
| 145 | Turning Our Custom Datasets Into DataLoaders | 07:00 | |
| 146 | Exploring State of the Art Data Augmentation With Torchvision Transforms | 14:24 | |
| 147 | Building a Baseline Model (Part 1): Loading and Transforming Data | 08:16 | |
| 148 | Building a Baseline Model (Part 2): Replicating Tiny VGG from Scratch | 11:25 | |
| 149 | Building a Baseline Model (Part 3): Doing a Forward Pass to Test Our Model Shapes | 08:10 | |
| 150 | Using the Torchinfo Package to Get a Summary of Our Model | 06:39 | |
| 151 | Creating Training and Testing loop Functions | 13:04 | |
| 152 | Creating a Train Function to Train and Evaluate Our Models | 10:15 | |
| 153 | Training and Evaluating Model 0 With Our Training Functions | 09:54 | |
| 154 | Plotting the Loss Curves of Model 0 | 09:03 | |
| 155 | Discussing the Balance Between Overfitting and Underfitting and How to Deal With Each | 14:14 | |
| 156 | Creating Augmented Training Datasets and DataLoaders for Model 1 | 11:04 | |
| 157 | Constructing and Training Model 1 | 07:11 | |
| 158 | Plotting the Loss Curves of Model 1 | 03:23 | |
| 159 | Plotting the Loss Curves of All of Our Models Against Each Other | 10:56 | |
| 160 | Predicting on Custom Data (Part 1): Downloading an Image | 05:33 | |
| 161 | Predicting on Custom Data (Part2): Loading In a Custom Image With PyTorch | 07:01 | |
| 162 | Predicting on Custom Data (Part 3): Getting Our Custom Image Into the Right Format | 14:08 | |
| 163 | Predicting on Custom Data (Part 4): Turning Our Models Raw Outputs Into Prediction Labels | 04:25 | |
| 164 | Predicting on Custom Data (Part 5): Putting It All Together | 12:48 | |
| 165 | Summary of What We Have Covered Plus Exercises and Extra-Curriculum | 06:05 | |
| 166 | What Is Going Modular and What We Are Going to Cover | 11:35 | |
| 167 | Going Modular Notebook (Part 1): Running It End to End | 07:41 | |
| 168 | Downloading a Dataset | 04:51 | |
| 169 | Writing the Outline for Our First Python Script to Setup the Data | 13:51 | |
| 170 | Creating a Python Script to Create Our PyTorch DataLoaders | 10:36 | |
| 171 | Turning Our Model Building Code into a Python Script | 09:19 | |
| 172 | Turning Our Model Training Code into a Python Script | 06:17 | |
| 173 | Turning Our Utility Function to Save a Model into a Python Script | 06:08 | |
| 174 | Creating a Training Script to Train Our Model in One Line of Code | 15:47 | |
| 175 | Going Modular: Summary, Exercises and Extra-Curriculum | 06:00 | |
| 176 | Introduction: What is Transfer Learning and Why Use It | 10:06 | |
| 177 | Where Can You Find Pretrained Models and What We Are Going to Cover | 05:13 | |
| 178 | Installing the Latest Versions of Torch and Torchvision | 08:06 | |
| 179 | Downloading Our Previously Written Code from Going Modular | 06:42 | |
| 180 | Downloading Pizza, Steak, Sushi Image Data from Github | 08:01 | |
| 181 | Turning Our Data into DataLoaders with Manually Created Transforms | 14:41 | |
| 182 | Turning Our Data into DataLoaders with Automatic Created Transforms | 13:07 | |
| 183 | Which Pretrained Model Should You Use | 12:16 | |
| 184 | Setting Up a Pretrained Model with Torchvision | 10:57 | |
| 185 | Different Kinds of Transfer Learning | 07:12 | |
| 186 | Getting a Summary of the Different Layers of Our Model | 06:50 | |
| 187 | Freezing the Base Layers of Our Model and Updating the Classifier Head | 13:27 | |
| 188 | Training Our First Transfer Learning Feature Extractor Model | 07:55 | |
| 189 | Plotting the Loss Curves of Our Transfer Learning Model | 06:27 | |
| 190 | Outlining the Steps to Make Predictions on the Test Images | 07:58 | |
| 191 | Creating a Function Predict On and Plot Images | 10:01 | |
| 192 | Making and Plotting Predictions on Test Images | 07:24 | |
| 193 | Making a Prediction on a Custom Image | 06:22 | |
| 194 | Main Takeaways, Exercises and Extra Curriculum | 03:22 | |
| 195 | What Is Experiment Tracking and Why Track Experiments | 07:07 | |
| 196 | Getting Setup by Importing Torch Libraries and Going Modular Code | 08:14 | |
| 197 | Creating a Function to Download Data | 10:24 | |
| 198 | Turning Our Data into DataLoaders Using Manual Transforms | 08:31 | |
| 199 | Turning Our Data into DataLoaders Using Automatic Transforms | 07:48 | |
| 200 | Preparing a Pretrained Model for Our Own Problem | 10:29 | |
| 201 | Setting Up a Way to Track a Single Model Experiment with TensorBoard | 13:36 | |
| 202 | Training a Single Model and Saving the Results to TensorBoard | 04:39 | |
| 203 | Exploring Our Single Models Results with TensorBoard | 10:18 | |
| 204 | Creating a Function to Create SummaryWriter Instances | 10:45 | |
| 205 | Adapting Our Train Function to Be Able to Track Multiple Experiments | 04:58 | |
| 206 | What Experiments Should You Try | 06:00 | |
| 207 | Discussing the Experiments We Are Going to Try | 06:02 | |
| 208 | Downloading Datasets for Our Modelling Experiments | 06:32 | |
| 209 | Turning Our Datasets into DataLoaders Ready for Experimentation | 08:29 | |
| 210 | Creating Functions to Prepare Our Feature Extractor Models | 15:55 | |
| 211 | Coding Out the Steps to Run a Series of Modelling Experiments | 14:28 | |
| 212 | Running Eight Different Modelling Experiments in 5 Minutes | 03:51 | |
| 213 | Viewing Our Modelling Experiments in TensorBoard | 13:39 | |
| 214 | Loading In the Best Model and Making Predictions on Random Images from the Test Set | 10:33 | |
| 215 | Making a Prediction on Our Own Custom Image with the Best Model | 03:45 | |
| 216 | Main Takeaways, Exercises and Extra Curriculum | 03:57 | |
| 217 | What Is a Machine Learning Research Paper? | 07:35 | |
| 218 | Why Replicate a Machine Learning Research Paper? | 03:14 | |
| 219 | Where Can You Find Machine Learning Research Papers and Code? | 08:19 | |
| 220 | What We Are Going to Cover | 08:22 | |
| 221 | Getting Setup for Coding in Google Colab | 08:22 | |
| 222 | Downloading Data for Food Vision Mini | 04:03 | |
| 223 | Turning Our Food Vision Mini Images into PyTorch DataLoaders | 09:48 | |
| 224 | Visualizing a Single Image | 03:46 | |
| 225 | Replicating a Vision Transformer - High Level Overview | 09:54 | |
| 226 | Breaking Down Figure 1 of the ViT Paper | 11:13 | |
| 227 | Breaking Down the Four Equations Overview and a Trick for Reading Papers | 10:56 | |
| 228 | Breaking Down Equation 1 | 08:15 | |
| 229 | Breaking Down Equations 2 and 3 | 10:04 | |
| 230 | Breaking Down Equation 4 | 07:28 | |
| 231 | Breaking Down Table 1 | 11:06 | |
| 232 | Calculating the Input and Output Shape of the Embedding Layer by Hand | 15:42 | |
| 233 | Turning a Single Image into Patches (Part 1: Patching the Top Row) | 15:04 | |
| 234 | Turning a Single Image into Patches (Part 2: Patching the Entire Image) | 12:34 | |
| 235 | Creating Patch Embeddings with a Convolutional Layer | 13:34 | |
| 236 | Exploring the Outputs of Our Convolutional Patch Embedding Layer | 12:55 | |
| 237 | Flattening Our Convolutional Feature Maps into a Sequence of Patch Embeddings | 10:00 | |
| 238 | Visualizing a Single Sequence Vector of Patch Embeddings | 05:04 | |
| 239 | Creating the Patch Embedding Layer with PyTorch | 17:02 | |
| 240 | Creating the Class Token Embedding | 13:25 | |
| 241 | Creating the Class Token Embedding - Less Birds | 13:25 | |
| 242 | Creating the Position Embedding | 11:26 | |
| 243 | Equation 1: Putting it All Together | 13:26 | |
| 244 | Equation 2: Multihead Attention Overview | 14:31 | |
| 245 | Equation 2: Layernorm Overview | 09:04 | |
| 246 | Turning Equation 2 into Code | 14:34 | |
| 247 | Checking the Inputs and Outputs of Equation | 05:41 | |
| 248 | Equation 3: Replication Overview | 09:12 | |
| 249 | Turning Equation 3 into Code | 11:26 | |
| 250 | Transformer Encoder Overview | 08:51 | |
| 251 | Combining Equation 2 and 3 to Create the Transformer Encoder | 09:17 | |
| 252 | Creating a Transformer Encoder Layer with In-Built PyTorch Layer | 15:55 | |
| 253 | Bringing Our Own Vision Transformer to Life - Part 1: Gathering the Pieces of the Puzzle | 18:20 | |
| 254 | Bringing Our Own Vision Transformer to Life - Part 2: Putting Together the Forward Method | 10:42 | |
| 255 | Getting a Visual Summary of Our Custom Vision Transformer | 07:14 | |
| 256 | Creating a Loss Function and Optimizer from the ViT Paper | 11:27 | |
| 257 | Training our Custom ViT on Food Vision Mini | 04:30 | |
| 258 | Discussing what Our Training Setup Is Missing | 09:09 | |
| 259 | Plotting a Loss Curve for Our ViT Model | 06:14 | |
| 260 | Getting a Pretrained Vision Transformer from Torchvision and Setting it Up | 14:38 | |
| 261 | Preparing Data to Be Used with a Pretrained ViT | 05:54 | |
| 262 | Training a Pretrained ViT Feature Extractor Model for Food Vision Mini | 07:16 | |
| 263 | Saving Our Pretrained ViT Model to File and Inspecting Its Size | 05:14 | |
| 264 | Discussing the Trade-Offs Between Using a Larger Model for Deployments | 03:47 | |
| 265 | Making Predictions on a Custom Image with Our Pretrained ViT | 03:31 | |
| 266 | PyTorch Paper Replicating: Main Takeaways, Exercises and Extra-Curriculum | 06:51 | |
| 267 | What is Machine Learning Model Deployment and Why Deploy a Machine Learning Model | 09:36 | |
| 268 | Three Questions to Ask for Machine Learning Model Deployment | 07:14 | |
| 269 | Where Is My Model Going to Go? | 13:35 | |
| 270 | How Is My Model Going to Function? | 08:00 | |
| 271 | Some Tools and Places to Deploy Machine Learning Models | 05:50 | |
| 272 | What We Are Going to Cover | 04:02 | |
| 273 | Getting Setup to Code | 06:16 | |
| 274 | Downloading a Dataset for Food Vision Mini | 03:24 | |
| 275 | Outlining Our Food Vision Mini Deployment Goals and Modelling Experiments | 08:00 | |
| 276 | Creating an EffNetB2 Feature Extractor Model | 09:46 | |
| 277 | Create a Function to Make an EffNetB2 Feature Extractor Model and Transforms | 06:30 | |
| 278 | Creating DataLoaders for EffNetB2 | 03:32 | |
| 279 | Training Our EffNetB2 Feature Extractor and Inspecting the Loss Curves | 09:16 | |
| 280 | Saving Our EffNetB2 Model to File | 03:25 | |
| 281 | Getting the Size of Our EffNetB2 Model in Megabytes | 05:52 | |
| 282 | Collecting Important Statistics and Performance Metrics for Our EffNetB2 Model | 06:35 | |
| 283 | Creating a Vision Transformer Feature Extractor Model | 07:52 | |
| 284 | Creating DataLoaders for Our ViT Feature Extractor Model | 02:31 | |
| 285 | Training Our ViT Feature Extractor Model and Inspecting Its Loss Curves | 06:20 | |
| 286 | Saving Our ViT Feature Extractor and Inspecting Its Size | 05:09 | |
| 287 | Collecting Stats About Our ViT Feature Extractor | 05:52 | |
| 288 | Outlining the Steps for Making and Timing Predictions for Our Models | 11:16 | |
| 289 | Creating a Function to Make and Time Predictions with Our Models | 16:21 | |
| 290 | Making and Timing Predictions with EffNetB2 | 10:44 | |
| 291 | Making and Timing Predictions with ViT | 07:35 | |
| 292 | Comparing EffNetB2 and ViT Model Statistics | 11:32 | |
| 293 | Visualizing the Performance vs Speed Trade-off | 15:55 | |
| 294 | Gradio Overview and Installation | 08:40 | |
| 295 | Gradio Function Outline | 08:50 | |
| 296 | Creating a Predict Function to Map Our Food Vision Mini Inputs to Outputs | 09:52 | |
| 297 | Creating a List of Examples to Pass to Our Gradio Demo | 05:27 | |
| 298 | Bringing Food Vision Mini to Life in a Live Web Application | 12:13 | |
| 299 | Getting Ready to Deploy Our App Hugging Face Spaces Overview | 06:27 | |
| 300 | Outlining the File Structure of Our Deployed App | 08:12 | |
| 301 | Creating a Food Vision Mini Demo Directory to House Our App Files | 04:12 | |
| 302 | Creating an Examples Directory with Example Food Vision Mini Images | 09:14 | |
| 303 | Writing Code to Move Our Saved EffNetB2 Model File | 07:43 | |
| 304 | Turning Our EffNetB2 Model Creation Function Into a Python Script | 04:02 | |
| 305 | Turning Our Food Vision Mini Demo App Into a Python Script | 13:28 | |
| 306 | Creating a Requirements File for Our Food Vision Mini App | 04:12 | |
| 307 | Downloading Our Food Vision Mini App Files from Google Colab | 11:31 | |
| 308 | Uploading Our Food Vision Mini App to Hugging Face Spaces Programmatically | 13:37 | |
| 309 | Running Food Vision Mini on Hugging Face Spaces and Trying it Out | 07:45 | |
| 310 | Food Vision Big Project Outline | 04:18 | |
| 311 | Preparing an EffNetB2 Feature Extractor Model for Food Vision Big | 09:39 | |
| 312 | Downloading the Food 101 Dataset | 07:46 | |
| 313 | Creating a Function to Split Our Food 101 Dataset into Smaller Portions | 13:37 | |
| 314 | Turning Our Food 101 Datasets into DataLoaders | 07:24 | |
| 315 | Training Food Vision Big: Our Biggest Model Yet! | 20:16 | |
| 316 | Outlining the File Structure for Our Food Vision Big | 05:49 | |
| 317 | Downloading an Example Image and Moving Our Food Vision Big Model File | 03:34 | |
| 318 | Saving Food 101 Class Names to a Text File and Reading them Back In | 06:57 | |
| 319 | Turning Our EffNetB2 Feature Extractor Creation Function into a Python Script | 02:21 | |
| 320 | Creating an App Script for Our Food Vision Big Model Gradio Demo | 10:42 | |
| 321 | Zipping and Downloading Our Food Vision Big App Files | 03:46 | |
| 322 | Deploying Food Vision Big to Hugging Face Spaces | 13:35 | |
| 323 | PyTorch Mode Deployment: Main Takeaways, Extra-Curriculum and Exercises | 06:14 | |
| 324 | Introduction to PyTorch 2.0 | 06:02 | |
| 325 | What We Are Going to Cover and PyTorch 2 Reference Materials | 01:22 | |
| 326 | Getting Started with PyTorch 2.0 in Google Colab | 04:20 | |
| 327 | PyTorch 2.0 - 30 Second Intro | 03:21 | |
| 328 | Getting Setup for PyTorch 2.0 | 02:23 | |
| 329 | Getting Info from Our GPUs and Seeing if They're Capable of Using PyTorch 2.0 | 06:50 | |
| 330 | Setting the Default Device in PyTorch 2.0 | 09:41 | |
| 331 | Discussing the Experiments We Are Going to Run for PyTorch 2.0 | 06:43 | |
| 332 | Creating a Function to Setup Our Model and Transforms | 10:18 | |
| 333 | Discussing How to Get Better Relative Speedups for Training Models | 08:24 | |
| 334 | Setting the Batch Size and Data Size Programmatically | 07:16 | |
| 335 | Getting More Speedups with TensorFloat-32 | 09:54 | |
| 336 | Downloading the CIFAR10 Dataset | 07:01 | |
| 337 | Creating Training and Test DataLoaders | 07:39 | |
| 338 | Preparing Training and Testing Loops with Timing Steps | 04:59 | |
| 339 | Experiment 1 - Single Run without Torch Compile | 08:23 | |
| 340 | Experiment 2 - Single Run with Torch Compile | 10:39 | |
| 341 | Comparing the Results of Experiments 1 and 2 | 11:20 | |
| 342 | Saving the Results of Experiments 1 and 2 | 04:40 | |
| 343 | Preparing Functions for Experiments 3 and 4 | 12:42 | |
| 344 | Experiment 3 - Training a Non-Compiled Model for Multiple Runs | 12:45 | |
| 345 | Experiment 4 - Training a Compiled Model for Multiple Runs | 09:58 | |
| 346 | Comparing the Results of Experiments 3 and 4 | 05:24 | |
| 347 | Potential Extensions and Resources to Learn More | 05:51 | |
| 348 | Thank You! | 01:18 |
Unlock unlimited learning
Get instant access to all 347 lessons in this course, plus thousands of other premium courses. One subscription, unlimited knowledge.
Learn more about subscriptionComments
0 commentsWant to join the conversation?
Sign in to commentSimilar courses
Machine Learning with Javascript
Sources: udemy, Stephen Grider
If you're here, you already know the truth: Machine Learning is the future of everything. In the coming years, there won't be a single industry in the world untouched by Machine...
17 hours 42 minutes 20 seconds
Analytics Engineering for Data Professionals
Sources: Fabrizio Valentini, Mattia Brunelli
Analytics Engineering is the foundation of Data Science and artificial intelligence. This approach represents a dynamic combination of data engineering and...
12 hours 46 minutes 13 seconds
Choosing Data Stores
Sources: Andreas Kretz
One of the key tasks when creating a data platform and pipelines is the selection of appropriate data storage systems. This course is dedicated to that topic.
1 hour 25 minutes 31 seconds
Data Engineering on Databricks
Sources: Andreas Kretz
Databricks is one of the most popular platforms for data processing using Apache Spark and creating modern data warehouses (Lakehouse).
1 hour 27 minutes 29 seconds
Machine Learning A-Z : Become Kaggle Master
Sources: udemy
Want to become a good Data Scientist? Then this is a right course for you. This course has been designed by IIT professionals who have mastered in Mathematics and Data Science....
36 hours 23 minutes 54 seconds