TensorFlow Developer Certificate in 2023: Zero to Mastery

62h 43m 54s
English
Paid
November 22, 2024

Learn TensorFlow. Pass the TensorFlow Developer Certificate Exam. Get Hired as a TensorFlow developer. This course will take you from a TensorFlow beginner to being part of Google's Certification Network.

More
  • Learn to pass Google's official TensorFlow Developer Certificate exam (and add it to your resume)
  • Complete access to ALL interactive notebooks and ALL course slides as downloadable guides
  • Understand how to integrate Machine Learning into tools and applications
  • Build image recognition, object detection, text recognition algorithms with deep neural networks and convolutional neural networks
  • Applying Deep Learning for Time Series Forecasting
  • Be recognized as a top candidate for recruiters seeking TensorFlow developers
  • Build TensorFlow models using Computer Vision, Convolutional Neural Networks and Natural Language Processing
  • Increase your skills in Machine Learning and Deep Learning
  • Learn to build all types of Machine Learning Models using the latest TensorFlow 2
  • Using real-world images in different shapes and sizes to visualize the journey of an image through convolutions to understand how a computer “sees” information, plot loss and accuracy
  • Gain the skills you need to become a TensorFlow Certified Developer

Watch Online TensorFlow Developer Certificate in 2023: Zero to Mastery

Join premium to watch
Go to premium
# Title Duration
1 Course Outline 05:22
2 What is deep learning? 04:39
3 Why use deep learning? 09:39
4 What are neural networks? 10:27
5 What is deep learning already being used for? 08:37
6 What is and why use TensorFlow? 07:57
7 What is a Tensor? 03:38
8 What we're going to cover throughout the course 04:30
9 How to approach this course 05:34
10 Creating your first tensors with TensorFlow and tf.constant() 18:46
11 Creating tensors with TensorFlow and tf.Variable() 07:08
12 Creating random tensors with TensorFlow 09:41
13 Shuffling the order of tensors 09:41
14 Creating tensors from NumPy arrays 11:56
15 Getting information from your tensors (tensor attributes 11:58
16 Indexing and expanding tensors 12:34
17 Manipulating tensors with basic operations 05:35
18 Matrix multiplication with tensors part 1 11:54
19 Matrix multiplication with tensors part 2 13:30
20 Matrix multiplication with tensors part 3 10:04
21 Changing the datatype of tensors 06:56
22 Tensor aggregation (finding the min, max, mean & more) 09:50
23 Tensor troubleshooting example (updating tensor datatypes) 06:14
24 Finding the positional minimum and maximum of a tensor (argmin and argmax) (9:31) 09:32
25 Squeezing a tensor (removing all 1-dimension axes) 03:00
26 One-hot encoding tensors 05:47
27 Trying out more tensor math operations 04:48
28 Exploring TensorFlow and NumPy's compatibility 05:44
29 Making sure our tensor operations run really fast on GPUs 10:20
30 Introduction to Neural Network Regression with TensorFlow 07:34
31 Inputs and outputs of a neural network regression model 09:00
32 Anatomy and architecture of a neural network regression model 07:56
33 Creating sample regression data (so we can model it) 12:47
34 The major steps in modelling with TensorFlow 20:16
35 Steps in improving a model with TensorFlow part 1 06:03
36 Steps in improving a model with TensorFlow part 2 09:26
37 Steps in improving a model with TensorFlow part 3 12:34
38 Evaluating a TensorFlow model part 1 ("visualise, visualise, visualise") 07:25
39 Evaluating a TensorFlow model part 2 (the three datasets) 11:02
40 Evaluating a TensorFlow model part 3 (getting a model summary) 17:19
41 Evaluating a TensorFlow model part 4 (visualising a model's layers) 07:15
42 Evaluating a TensorFlow model part 5 (visualising a model's predictions) 09:17
43 Evaluating a TensorFlow model part 6 (common regression evaluation metrics) 08:06
44 Evaluating a TensorFlow regression model part 7 (mean absolute error) 05:53
45 Evaluating a TensorFlow regression model part 7 (mean square error) 03:19
46 Setting up TensorFlow modelling experiments part 1 (start with a simple model) 13:51
47 Setting up TensorFlow modelling experiments part 2 (increasing complexity) 11:30
48 Comparing and tracking your TensorFlow modelling experiments 10:21
49 How to save a TensorFlow model 08:20
50 How to load and use a saved TensorFlow model 10:16
51 (Optional) How to save and download files from Google Colab 06:19
52 Putting together what we've learned part 1 (preparing a dataset) 13:32
53 Putting together what we've learned part 2 (building a regression model) 13:21
54 Putting together what we've learned part 3 (improving our regression model) 15:48
55 Preprocessing data with feature scaling part 1 (what is feature scaling?) 09:35
56 Preprocessing data with feature scaling part 2 (normalising our data) 10:58
57 Preprocessing data with feature scaling part 3 (fitting a model on scaled data) 07:41
58 Introduction to neural network classification in TensorFlow 08:26
59 Example classification problems (and their inputs and outputs) 06:39
60 Input and output tensors of classification problems 06:22
61 Typical architecture of neural network classification models with TensorFlow 09:37
62 Creating and viewing classification data to model 11:35
63 Checking the input and output shapes of our classification data 04:39
64 Building a not very good classification model with TensorFlow 12:11
65 Trying to improve our not very good classification model 09:14
66 Creating a function to view our model's not so good predictions 15:09
67 Make our poor classification model work for a regression dataset 12:19
68 Non-linearity part 1: Straight lines and non-straight lines 09:39
69 Non-linearity part 2: Building our first neural network with non-linearity 05:48
70 Non-linearity part 3: Upgrading our non-linear model with more layers 10:19
71 Non-linearity part 4: Modelling our non-linear data once and for all 08:38
72 Non-linearity part 5: Replicating non-linear activation functions from scratch 14:27
73 Getting great results in less time by tweaking the learning rate 14:48
74 Using the TensorFlow History object to plot a model's loss curves 06:12
75 Using callbacks to find a model's ideal learning rate 17:33
76 Training and evaluating a model with an ideal learning rate 09:21
77 Introducing more classification evaluation methods 06:05
78 Finding the accuracy of our classification model 04:18
79 Creating our first confusion matrix (to see where our model is getting confused) 08:28
80 Making our confusion matrix prettier 14:01
81 Putting things together with multi-class classification part 1: Getting the data 10:38
82 Multi-class classification part 2: Becoming one with the data 07:08
83 Multi-class classification part 3: Building a multi-class classification model 15:39
84 Multi-class classification part 4: Improving performance with normalisation 12:44
85 Multi-class classification part 5: Comparing normalised and non-normalised data 04:14
86 Multi-class classification part 6: Finding the ideal learning rate 10:39
87 Multi-class classification part 7: Evaluating our model 13:17
88 Multi-class classification part 8: Creating a confusion matrix 04:27
89 Multi-class classification part 9: Visualising random model predictions 10:43
90 What "patterns" is our model learning? 15:34
91 Introduction to Computer Vision with TensorFlow 09:37
92 Introduction to Convolutional Neural Networks (CNNs) with TensorFlow 08:00
93 Downloading an image dataset for our first Food Vision model 08:28
94 Becoming One With Data 05:06
95 Becoming One With Data Part 2 12:27
96 Becoming One With Data Part 3 04:23
97 Building an end to end CNN Model 18:18
98 Using a GPU to run our CNN model 5x faster 09:18
99 Trying a non-CNN model on our image data 08:52
100 Improving our non-CNN model by adding more layers 09:53
101 Breaking our CNN model down part 1: Becoming one with the data 09:04
102 Breaking our CNN model down part 2: Preparing to load our data 11:47
103 Breaking our CNN model down part 3: Loading our data with ImageDataGenerator 09:55
104 Breaking our CNN model down part 4: Building a baseline CNN model 08:03
105 Breaking our CNN model down part 5: Looking inside a Conv2D layer 15:21
106 Breaking our CNN model down part 6: Compiling and fitting our baseline CNN 07:15
107 Breaking our CNN model down part 7: Evaluating our CNN's training curves 11:46
108 Breaking our CNN model down part 8: Reducing overfitting with Max Pooling 13:41
109 Breaking our CNN model down part 9: Reducing overfitting with data augmentation 06:53
110 Breaking our CNN model down part 10: Visualizing our augmented data 15:05
111 Breaking our CNN model down part 11: Training a CNN model on augmented data 08:50
112 Breaking our CNN model down part 12: Discovering the power of shuffling data 10:02
113 Breaking our CNN model down part 13: Exploring options to improve our model 05:22
114 Downloading a custom image to make predictions on 04:55
115 Writing a helper function to load and preprocessing custom images 10:01
116 Making a prediction on a custom image with our trained CNN 10:09
117 Multi-class CNN's part 1: Becoming one with the data 15:00
118 Multi-class CNN's part 2: Preparing our data (turning it into tensors) 06:39
119 Multi-class CNN's part 3: Building a multi-class CNN model 07:25
120 Multi-class CNN's part 4: Fitting a multi-class CNN model to the data 06:03
121 Multi-class CNN's part 5: Evaluating our multi-class CNN model ( 04:52
122 Multi-class CNN's part 6: Trying to fix overfitting by removing layers 12:20
123 Multi-class CNN's part 7: Trying to fix overfitting with data augmentation 11:47
124 Multi-class CNN's part 8: Things you could do to improve your CNN model 04:24
125 Multi-class CNN's part 9: Making predictions with our model on custom images 09:23
126 Saving and loading our trained CNN model 06:22
127 What is and why use transfer learning? 10:13
128 Downloading and preparing data for our first transfer learning model 14:40
129 Introducing Callbacks in TensorFlow and making a callback to track our models 10:02
130 Exploring the TensorFlow Hub website for pretrained models 09:52
131 Building and compiling a TensorFlow Hub feature extraction model 14:01
132 Blowing our previous models out of the water with transfer learning 09:14
133 Plotting the loss curves of our ResNet feature extraction model 07:36
134 Building and training a pre-trained EfficientNet model on our data 09:43
135 Different Types of Transfer Learning 11:41
136 Comparing Our Model's Results 15:17
137 Introduction to Transfer Learning in TensorFlow Part 2: Fine-tuning 06:17
138 Importing a script full of helper functions (and saving lots of space) 07:36
139 Downloading and turning our images into a TensorFlow BatchDataset 15:39
140 Discussing the four (actually five) modelling experiments we're running 02:16
141 Comparing the TensorFlow Keras Sequential API versus the Functional API 02:35
142 Creating our first model with the TensorFlow Keras Functional API 11:39
143 Compiling and fitting our first Functional API model 10:54
144 Getting a feature vector from our trained model 13:40
145 Drilling into the concept of a feature vector (a learned representation) 03:44
146 Downloading and preparing the data for Model 1 (1 percent of training data) 09:52
147 Building a data augmentation layer to use inside our model 12:07
148 Visualising what happens when images pass through our data augmentation layer 10:56
149 Building Model 1 (with a data augmentation layer and 1% of training data) 15:56
150 Building Model 2 (with a data augmentation layer and 10% of training data) 16:38
151 Creating a ModelCheckpoint to save our model's weights during training 07:26
152 Fitting and evaluating Model 2 (and saving its weights using ModelCheckpoint) 07:15
153 Loading and comparing saved weights to our existing trained Model 2 07:18
154 Preparing Model 3 (our first fine-tuned model) 20:27
155 Fitting and evaluating Model 3 (our first fine-tuned model) 07:46
156 Comparing our model's results before and after fine-tuning 10:27
157 Downloading and preparing data for our biggest experiment yet (Model 4) 06:25
158 Preparing our final modelling experiment (Model 4) 12:01
159 Fine-tuning Model 4 on 100% of the training data and evaluating its results 10:20
160 Comparing our modelling experiment results in TensorBoard 10:47
161 How to view and delete previous TensorBoard experiments 02:05
162 Introduction to Transfer Learning Part 3: Scaling Up 06:20
163 Getting helper functions ready and downloading data to model 13:35
164 Outlining the model we're going to build and building a ModelCheckpoint callback 05:39
165 Creating a data augmentation layer to use with our model 04:40
166 Creating a headless EfficientNetB0 model with data augmentation built in 08:59
167 Fitting and evaluating our biggest transfer learning model yet 07:57
168 Unfreezing some layers in our base model to prepare for fine-tuning 11:29
169 Fine-tuning our feature extraction model and evaluating its performance 08:24
170 Saving and loading our trained model 06:26
171 Downloading a pretrained model to make and evaluate predictions with 06:35
172 Making predictions with our trained model on 25,250 test samples 12:47
173 Unravelling our test dataset for comparing ground truth labels to predictions 06:06
174 Confirming our model's predictions are in the same order as the test labels 05:18
175 Creating a confusion matrix for our model's 101 different classes 12:08
176 Evaluating every individual class in our dataset 14:17
177 Plotting our model's F1-scores for each separate class 07:37
178 Creating a function to load and prepare images for making predictions 12:09
179 Making predictions on our test images and evaluating them 16:07
180 Discussing the benefits of finding your model's most wrong predictions 06:10
181 Writing code to uncover our model's most wrong predictions 11:17
182 Plotting and visualizing the samples our model got most wrong 10:37
183 Making predictions on and plotting our own custom images 09:50
184 Introduction to Milestone Project 1: Food Vision Big™ 05:45
185 Making sure we have access to the right GPU for mixed precision training 10:18
186 Getting helper functions ready 03:07
187 Introduction to TensorFlow Datasets (TFDS) 12:04
188 Exploring and becoming one with the data (Food101 from TensorFlow Datasets) 15:57
189 Creating a preprocessing function to prepare our data for modelling 15:51
190 Batching and preparing our datasets (to make them run fast) 13:48
191 Exploring what happens when we batch and prefetch our data 06:50
192 Creating modelling callbacks for our feature extraction model 07:15
193 Turning on mixed precision training with TensorFlow 10:06
194 Creating a feature extraction model capable of using mixed precision training 12:43
195 Checking to see if our model is using mixed precision training layer by layer 07:57
196 Training and evaluating a feature extraction model (Food Vision Big™) 10:20
197 Introducing your Milestone Project 1 challenge: build a model to beat DeepFood 07:48
198 Introduction to Natural Language Processing (NLP) and Sequence Problems 12:52
199 Example NLP inputs and outputs 07:23
200 The typical architecture of a Recurrent Neural Network (RNN) 09:04
201 Preparing a notebook for our first NLP with TensorFlow project 08:53
202 Becoming one with the data and visualizing a text dataset 16:42
203 Splitting data into training and validation sets 06:27
204 Converting text data to numbers using tokenisation and embeddings (overview) 09:23
205 Setting up a TensorFlow TextVectorization layer to convert text to numbers 17:11
206 Mapping the TextVectorization layer to text data and turning it into numbers 11:03
207 Creating an Embedding layer to turn tokenised text into embedding vectors 12:28
208 Discussing the various modelling experiments we're going to run 08:58
209 Model 0: Building a baseline model to try and improve upon 09:26
210 Creating a function to track and evaluate our model's results 12:15
211 Model 1: Building, fitting and evaluating our first deep model on text data 20:52
212 Visualizing our model's learned word embeddings with TensorFlow's projector tool 20:44
213 High-level overview of Recurrent Neural Networks (RNNs) + where to learn more 09:35
214 Model 2: Building, fitting and evaluating our first TensorFlow RNN model (LSTM) 18:17
215 Model 3: Building, fitting and evaluating a GRU-cell powered RNN 16:57
216 Model 4: Building, fitting and evaluating a bidirectional RNN model 19:35
217 Discussing the intuition behind Conv1D neural networks for text and sequences 19:32
218 Model 5: Building, fitting and evaluating a 1D CNN for text 09:58
219 Using TensorFlow Hub for pretrained word embeddings (transfer learning for NLP) 13:46
220 Model 6: Building, training and evaluating a transfer learning model for NLP 10:46
221 Preparing subsets of data for model 7 (same as model 6 but 10% of data) 10:53
222 Model 7: Building, training and evaluating a transfer learning model on 10% data 10:05
223 Fixing our data leakage issue with model 7 and retraining it 13:43
224 Comparing all our modelling experiments evaluation metrics 13:15
225 Uploading our model's training logs to TensorBoard and comparing them 11:15
226 Saving and loading in a trained NLP model with TensorFlow 10:26
227 Downloading a pretrained model and preparing data to investigate predictions 13:25
228 Visualizing our model's most wrong predictions 08:29
229 Making and visualizing predictions on the test dataset 08:28
230 Understanding the concept of the speed/score tradeoff 15:02
231 Introduction to Milestone Project 2: SkimLit 14:21
232 What we're going to cover in Milestone Project 2 (NLP for medical abstracts) 07:23
233 SkimLit inputs and outputs 11:03
234 Setting up our notebook for Milestone Project 2 (getting the data) 14:59
235 Visualizing examples from the dataset (becoming one with the data) 13:19
236 Writing a preprocessing function to structure our data for modelling 19:51
237 Performing visual data analysis on our preprocessed text 07:56
238 Turning our target labels into numbers (ML models require numbers) 13:16
239 Model 0: Creating, fitting and evaluating a baseline model for SkimLit 09:26
240 Preparing our data for deep sequence models 09:56
241 Creating a text vectoriser to map our tokens (text) to numbers 14:08
242 Creating a custom token embedding layer with TensorFlow 09:15
243 Creating fast loading dataset with the TensorFlow tf.data API 09:50
244 Model 1: Building, fitting and evaluating a Conv1D with token embeddings 17:22
245 Preparing a pretrained embedding layer from TensorFlow Hub for Model 2 10:54
246 Model 2: Building, fitting and evaluating a Conv1D model with token embeddings 11:31
247 Creating a character-level tokeniser with TensorFlow's TextVectorization layer 23:25
248 Creating a character-level embedding layer with tf.keras.layers.Embedding 07:45
249 Model 3: Building, fitting and evaluating a Conv1D model on character embeddings 13:46
250 Discussing how we're going to build Model 4 (character + token embeddings) 06:05
251 Model 4: Building a multi-input model (hybrid token + character embeddings) 15:37
252 Model 4: Plotting and visually exploring different data inputs 07:33
253 Crafting multi-input fast loading tf.data datasets for Model 4 08:42
254 Model 4: Building, fitting and evaluating a hybrid embedding model 13:19
255 Model 5: Adding positional embeddings via feature engineering (overview) 07:19
256 Encoding the line number feature to used with Model 5 12:26
257 Encoding the total lines feature to be used with Model 5 07:57
258 Model 5: Building the foundations of a tribrid embedding model 09:20
259 Model 5: Completing the build of a tribrid embedding model for sequences 14:09
260 Visually inspecting the architecture of our tribrid embedding model 10:26
261 Creating multi-level data input pipelines for Model 5 with the tf.data API 09:01
262 Bringing SkimLit to life!!! (fitting and evaluating Model 5) 10:36
263 Comparing the performance of all of our modelling experiments 09:37
264 Saving, loading & testing our best performing model 07:49
265 Congratulations and your challenge before heading to the next module 12:34
266 Introduction to Milestone Project 3 (BitPredict) & where you can get help 03:54
267 What is a time series problem and example forecasting problems at Uber 07:47
268 Example forecasting problems in daily life 04:53
269 What can be forecast? 07:58
270 What we're going to cover (broadly) 02:36
271 Time series forecasting inputs and outputs 08:56
272 Downloading and inspecting our Bitcoin historical dataset 14:59
273 Different kinds of time series patterns & different amounts of feature variables 07:40
274 Visualizing our Bitcoin historical data with pandas 04:53
275 Reading in our Bitcoin data with Python's CSV module 10:59
276 Creating train and test splits for time series (the wrong way) 08:38
277 Creating train and test splits for time series (the right way) 07:13
278 Creating a plotting function to visualize our time series data 07:58
279 Discussing the various modelling experiments were going to be running 09:12
280 Model 0: Making and visualizing a naive forecast model 12:17
281 Discussing some of the most common time series evaluation metrics 11:12
282 Implementing MASE with TensorFlow 09:39
283 Creating a function to evaluate our model's forecasts with various metrics 10:12
284 Discussing other non-TensorFlow kinds of time series forecasting models 05:07
285 Formatting data Part 2: Creating a function to label our windowed time series 13:02
286 Discussing the use of windows and horizons in time series data 07:51
287 Writing a preprocessing function to turn time series data into windows & labels 23:36
288 Turning our windowed time series data into training and test sets 10:02
289 Creating a modelling checkpoint callback to save our best performing model 07:26
290 Model 1: Building, compiling and fitting a deep learning model on Bitcoin data 16:59
291 Creating a function to make predictions with our trained models 14:04
292 Model 2: Building, fitting and evaluating a deep model with a larger window size-27 17:44
293 Model 3: Building, fitting and evaluating a model with a larger horizon size 13:16
294 Adjusting the evaluation function to work for predictions with larger horizons 08:35
295 Model 3: Visualizing the results 08:45
296 Comparing our modelling experiments so far and discussing autocorrelation 09:45
297 Preparing data for building a Conv1D model 13:22
298 Model 4: Building, fitting and evaluating a Conv1D model on our Bitcoin data 14:52
299 Model 5: Building, fitting and evaluating a LSTM (RNN) model on our Bitcoin data 16:06
300 Investigating how to turn our univariate time series into multivariate 13:53
301 Creating and plotting a multivariate time series with BTC price and block reward 12:13
302 Preparing our multivariate time series for a model 13:38
303 Model 6: Building, fitting and evaluating a multivariate time series model 09:26
304 Model 7: Discussing what we're going to be doing with the N-BEATS algorithm 09:40
305 Model 7: Replicating the N-BEATS basic block with TensorFlow layer subclassing 18:39
306 Model 7: Testing our N-BEATS block implementation with dummy data inputs 15:03
307 Model 7: Setting up hyperparameters for the N-BEATS algorithm 08:51
308 Model 7: Getting ready for residual connections 12:56
309 Model 7: Outlining the steps we're going to take to build the N-BEATS model 10:06
310 Model 7: Putting together the pieces of the puzzle of the N-BEATS model 22:23
311 Model 7: Plotting the N-BEATS algorithm we've created and admiring its beauty 06:47
312 Model 8: Ensemble model overview 04:44
313 Model 8: Building, compiling and fitting an ensemble of models 20:05
314 Model 8: Making and evaluating predictions with our ensemble model 16:10
315 Discussing the importance of prediction intervals in forecasting 12:57
316 Getting the upper and lower bounds of our prediction intervals 07:58
317 Plotting the prediction intervals of our ensemble model predictions 13:03
318 (Optional) Discussing the types of uncertainty in machine learning 13:42
319 Model 9: Preparing data to create a model capable of predicting into the future 08:25
320 Model 9: Building, compiling and fitting a future predictions model 05:02
321 Model 9: Discussing what's required for our model to make future predictions 08:31
322 Model 9: Creating a function to make forecasts into the future 12:09
323 Model 9: Plotting our model's future forecasts 13:10
324 Model 10: Introducing the turkey problem and making data for it 14:16
325 Model 10: Building a model to predict on turkey data (why forecasting is BS) 13:39
326 Comparing the results of all of our models and discussing where to go next 13:00
327 What is the TensorFlow Developer Certification? 05:29
328 Why the TensorFlow Developer Certification? 06:58
329 How to prepare (your brain) for the TensorFlow Developer Certification 08:15
330 How to prepare (your computer) for the TensorFlow Developer Certification 12:44
331 What to do after the TensorFlow Developer Certification exam 02:14
332 What is Machine Learning? 04:52
333 AI/Machine Learning/Data Science 06:53
334 Exercise: Machine Learning Playground 06:17
335 How Did We Get Here? 06:04
336 Exercise: YouTube Recommendation Engine 04:25
337 Types of Machine Learning 04:42
338 What Is Machine Learning? Round 2 04:45
339 Section Review 01:49
340 Section Overview 02:39
341 Introducing Our Framework 03:09
342 6 Step Machine Learning Framework 05:00
343 Types of Machine Learning Problems 10:33
344 Types of Data 04:51
345 Types of Evaluation 03:32
346 Features In Data 05:59
347 Modelling - Splitting Data 05:23
348 Modelling - Picking the Model 04:36
349 Modelling - Tuning 03:18
350 Modelling - Comparison 03:36
351 Experimentation 09:33
352 Tools We Will Use 04:01
353 Section Overview 02:28
354 Pandas Introduction 04:30
355 Series, Data Frames and CSVs 13:22
356 Describing Data with Pandas 09:49
357 Selecting and Viewing Data with Pandas 11:09
358 Selecting and Viewing Data with Pandas Part 2 13:07
359 Manipulating Data 13:57
360 Manipulating Data 2 09:57
361 Manipulating Data 3 10:13
362 How To Download The Course Assignments 02:41
363 Section Overview 07:44
364 NumPy Introduction 05:18
365 NumPy DataTypes and Attributes 14:06
366 Creating NumPy Arrays 09:23
367 NumPy Random Seed 07:18
368 Viewing Arrays and Matrices 09:36
369 Manipulating Arrays 11:32
370 Manipulating Arrays 2 09:45
371 Standard Deviation and Variance 07:11
372 Reshape and Transpose 07:27
373 Dot Product vs Element Wise 11:46
374 Exercise: Nut Butter Store Sales 03:34
375 Comparison Operators 13:05
376 Sorting Arrays 06:20
377 Turn Images Into NumPy Arrays 07:38

Similar courses to TensorFlow Developer Certificate in 2023: Zero to Mastery

PyTorch for Deep Learning and Computer Vision

PyTorch for Deep Learning and Computer Visionudemy

Duration 10 hours 20 minutes 51 seconds
The Data Bootcamp: Transform your Data using dbt™

The Data Bootcamp: Transform your Data using dbt™udemy

Duration 4 hours 10 minutes 51 seconds
PyTorch for Deep Learning

PyTorch for Deep Learningzerotomastery.io

Duration 52 hours 27 seconds
Data Analysis with Pandas and Python

Data Analysis with Pandas and Pythonudemy

Duration 19 hours 5 minutes 40 seconds
Time Series Analysis, Forecasting, and Machine Learning

Time Series Analysis, Forecasting, and Machine Learningudemy

Duration 22 hours 47 minutes 45 seconds
Complete linear algebra: theory and implementation

Complete linear algebra: theory and implementationudemy

Duration 32 hours 53 minutes 26 seconds