GCPAMLTF - Advanced Machine Learning with TensorFlow on Google Cloud Platform

I.       Overview:

This training focuses on advanced machine learning topics using Google Cloud Platform where you will get hands-on experience optimizing, deploying, and scaling production ML models of various types in hands-on labs. This training picks up where “Machine Learning on Google Cloud Platform” left off and teaches you how to build scalable, accurate, and production-ready models for structured data, image data, time-series, and natural language text. It ends with a module on building recommendation systems.

II.     Duration: 05 days

III. Objectives

  • Implement the various flavors of production ML systems—static, dynamic, and continuous training; static and dynamic inference; and batch and online processing.
  • Solve an ML problem by building an end-to-end pipeline, going from data exploration, preprocessing, feature engineering, model building, hyperparameter tuning, deployment, and serving.
  • Develop a range of image classification models from simple linear models to high-performing convolutional neural networks (CNNs) with batch normalization, augmentation, and transfer learning.
  • Forecast time-series values using CNNs, recurrent neural networks (RNNs), and LSTMs.
  • Apply ML to natural language text using CNNs, RNNs, LSTMs, reusable word embeddings, and encoder-decoder generative models.
  • Implement content-based, collaborative, hybrid, and neural recommendation models in TensorFlow.

IV. Intended Audience

  • Data Engineers and programmers interested in learning how to apply machine learning in practice.
  • Anyone interested in learning how to leverage machine learning in their enterprise.

V.    Prerequisites

To get the most out of this training, participants should have:

  • Knowledge of machine learning and TensorFlow to the level covered in Machine Learning on Google Cloud Platform coursework.
  • Experience coding in Python.
  • Knowledge of basic statistics.
  •  Knowledge of SQL and cloud computing (helpful).

VI.  Outline

Module 1: Machine Learning on Google Cloud Platform

  •  Effective ML.
  •  Fully managed ML.

Module 2: Explore the Data

  •  Exploring the dataset.
  •  BigQuery.
  •  BigQuery and AI Platform Notebooks.

Module 3: Creating the dataset

  •  Creating a dataset.
  • Module 4: Build the Model
  •  Build the model.

Module 5: Operationalize the model

  •  Operationalizing the model.
  •  Cloud AI Platform.
  •  Train and deploy with Cloud AI Platform.
  •  BigQuery ML.
  •  Deploying and Predicting with Cloud AI Platform.

Module 6: Architecting Production ML Systems

  •  The Components of an ML System.
  •  The Components of an ML System: Data Analysis and Validation.
  •  The Components of an ML System: Data Transformation + Trainer.
  •  The Components of an ML System: Tuner + Model Evaluation and Validation.
  •  The Components of an ML System: Serving.
  •  The Components of an ML System: Orchestration + Workflow.
  •  The Components of an ML System: Integrated Frontend + Storage.
  •  Training Design Decisions.
  •  Serving Design Decisions.
  •  Designing from Scratch.

Module 7: Ingesting data for Cloud-based analytics and ML

  •  Data On-Premise.
  •  Large Datasets.
  •  Data on Other Clouds.
  •  Existing Databases.

Module 8: Designing Adaptable ML systems

  •  Adapting to Data.
  •  Changing Distributions.
  •  Right and Wrong Decisions.
  •  System Failure.
  •  Mitigating Training-Serving Skew through Design.
  •  Debugging a Production Model.

Module 9: Designing High-performance ML systems

  •  Training.
  •  Predictions.
  •  Why distributed training?
  •  Distributed training architectures.
  •  Faster input pipelines.
  •  Native TensorFlow Operations.
  •  TensorFlow Records.
  •  Parallel pipelines.
  •  Data parallelism with All Reduce.
  •  Parameter Server Approach.
  •  Inference.

Module 10: Hybrid ML systems

  •  Machine Learning on Hybrid Cloud.
  •  KubeFlow.
  •  Embedded Models.
  •  TensorFlow Lite.
  •  Optimizing for Mobile.

Module 11: Welcome to Image Understanding with TensorFlow on GCP

  •  Images as Visual Data.
  •  Structured vs Unstructured Data.

Module 12: Linear and DNN Models

  •  Linear Models.
  •  DNN Models Review.
  •  Review: What is Dropout?

Module 13: Convolutional Neural Networks (CNNs)

  •  Understanding Convolutions.
  •  CNN Model Parameters.
  •  Working with Pooling Layers.
  •  Implementing CNNs with TensorFlow.

Module 14: Dealing with Data Scarcity

  •  The Data Scarcity Problem.
  •  Data Augmentation.
  •  Transfer Learning.
  •  No Data, No Problem.

Module 15: Going Deeper Faster

  •  Batch Normalization.
  •  Residual Networks.
  •  Accelerators (CPU vs GPU, TPU).
  •  TPU Estimator.
  •  Neural Architecture Search.

Module 16: Pre-built ML Models for Image Classification

  •  Pre-built ML Models.
  •  Cloud Vision API.
  •  AutoML Vision.
  •  AutoML Architecture.

Module 17: Working with Sequences

  •  Sequence data and models.
  •  From sequences to inputs,
  •  Modeling sequences with linear models.
  •  Modeling sequences with DNNs.
  •  Modeling sequences with CNNs.
  •  The variable-length problem4m.

Module 18: Recurrent Neural Networks

  •  Introducing Recurrent Neural Networks.
  •  How RNNs represent the past.
  •  The limits of what RNNs can represent.
  •  The vanishing gradient problem.

Module 19: Dealing with Longer Sequences

  •  LSTMs and GRUs.
  •  RNNs in TensorFlow.
  •  Deep RNNs.
  •  Improving our Loss Function.
  •  Working with Real Data.

Module 20: Text Classification

  •  Working with Text.
  •  Text Classification.
  •  Selecting a Model.
  •  Python vs Native TensorFlow.

Module 21: Reusable Embeddings

  •  Historical methods of making word embeddings.
  •  Modern methods of making word embeddings.
  •  Introducing TensorFlow Hub.
  •  Using TensorFlow Hub within an estimator.

Module 22: Recurrent Neural NetworksEncoder-Decoder Models

  •  Introducing Encoder-Decoder Networks.
  •  Attention Networks.
  •  Training Encoder-Decoder Models with TensorFlow.
  •  Introducing Tensor2Tensor.
  •  AutoML Translation.
  •  Dialogflow.

Module 23: Recommendation Systems Overview

  •  Types of Recommendation Systems.
  •  Content-Based or Collaborative.
  •  Recommendation System Pitfalls.

Module 24:Content-Based Recommendation Systems

  •  Content-Based Recommendation Systems.
  •  Similarity Measures.
  •  Building a User Vector.
  •  Making Recommendations Using a User Vector.
  •  Making Recommendations for Many Users.
  •  Using Neural Networks for Content-Based Recommendation Systems.

Module 25:Collaborative Filtering Recommendation Systems

  •  Types of User Feedback Data.
  •  Embedding Users and Items.
  •  Factorization Approaches.
  •  The ALS Algorithm.
  •  Preparing Input Data for ALS.
  •  Creating Sparse Tensors For Efficient WALS Input.
  •  Instantiating a WALS Estimator: From Input to Estimator.
  •  Instantiating a WAL Estimator: Decoding TFRecords.
  •  Instantiating a WALS Estimator: Recovering Keys.
  •  Instantiating a WALS Estimator: Training and Prediction.
  •  Issues with Collaborative Filtering.
  •  Cold Starts.

Module 26:Neural Networks for Recommendation Systems

  •  Hybrid Recommendation System.
  •  Context-Aware Recommendation Systems.
  •  Context-Aware Algorithms.
  •  Contextual Postfiltering.
  •  Modeling Using Context-Aware Algorithms.

Module 27:Building an End-to-End Recommendation System

  •  Architecture Overview.
  •  Cloud Composer Overview.
  •  Cloud Composer: DAGs.
  •  Cloud Composer: Operators for ML9.
  •  Cloud Composer: Scheduling.
  •  Cloud Composer: Triggering Workflows with Cloud Functions.
  •  Cloud Composer: Monitoring and Logging.
  • On demand
  • Take this course on demand
  • Upcoming classrooms
  • There are no upcoming instructor-led sessions
  • Học trực tuyến

  • Học tại Hồ Chí Minh

  • Học tại Hà Nội


Các khóa học khác