Podcast
Questions and Answers
You are building an ML model to detect anomalies in real-time sensor data. You will use Pub/Sub to handle incoming
requests. You want to store the results for analytics and visualization. How should you configure the pipeline?
You are building an ML model to detect anomalies in real-time sensor data. You will use Pub/Sub to handle incoming requests. You want to store the results for analytics and visualization. How should you configure the pipeline?
- 1 = Dataflow, 2 = AI Platform, 3 = BigQuery (correct)
- 1 = DataProc, 2 = AutoML, 3 = Cloud Bigtable
- 1 = BigQuery, 2 = AutoML, 3 = Cloud Functions
- 1 = BigQuery, 2 = AI Platform, 3 = Cloud Storage
Your team needs to build a model that predicts whether images contain a driver's license, passport, or credit card. The data engineering team already built the pipeline and generated a dataset composed of 10,000 images with driver's licenses, 1,000 images with passports, and 1,000 images with credit cards. You now have to train a model with the following label map: [Ëœdrivers_license',
Ëœpassport', `Ëœcredit_card']. Which loss function should you use?
Your team needs to build a model that predicts whether images contain a driver's license, passport, or credit card. The data engineering team already built the pipeline and generated a dataset composed of 10,000 images with driver's licenses, 1,000 images with passports, and 1,000 images with credit cards. You now have to train a model with the following label map: [Ëœdrivers_license',
Ëœpassport', `Ëœcredit_card']. Which loss function should you use?
- Categorical hinge
- Binary cross-entropy
- Sparse categorical cross-entropy
- Categorical cross-entropy (correct)
You are an ML engineer at a manufacturing company. You need to build a model that identifies defects in products
based on images of the product taken at the end of the assembly line. You want your model to preprocess the
images with lower computation to quickly extract features of defects in products. Which approach should you use
to build the model?
You are an ML engineer at a manufacturing company. You need to build a model that identifies defects in products based on images of the product taken at the end of the assembly line. You want your model to preprocess the images with lower computation to quickly extract features of defects in products. Which approach should you use to build the model?
- Reinforcement learning
- Recommender system
- Recurrent Neural Networks (RNN)
- Convolutional Neural Networks (CNN) (correct)
You are developing an ML model intended to classify whether X-ray images indicate bone fracture risk. You have trained a ResNet architecture on Vertex AI using a TPU as an accelerator, however you are unsatisfied with the training time and memory usage. You want to quickly iterate your training code but make minimal changes to the code. You also want to minimize impact on the model’s accuracy. What should you do?
You are developing an ML model intended to classify whether X-ray images indicate bone fracture risk. You have trained a ResNet architecture on Vertex AI using a TPU as an accelerator, however you are unsatisfied with the training time and memory usage. You want to quickly iterate your training code but make minimal changes to the code. You also want to minimize impact on the model’s accuracy. What should you do?
You have successfully deployed to production a large and complex TensorFlow model trained on tabular data. You want to predict the lifetime value (LTV) field for each subscription stored in the BigQuery table named subscription. subscriptionPurchase in the project named my-fortune500-company-project. You have organized all your training code, from preprocessing data from the BigQuery table up to deploying the validated model to the Vertex AI endpoint, into a TensorFlow Extended (TFX) pipeline. You want to prevent prediction drift, i.e., a situation when a feature data distribution in production changes significantly over time. What should you do?
You have successfully deployed to production a large and complex TensorFlow model trained on tabular data. You want to predict the lifetime value (LTV) field for each subscription stored in the BigQuery table named subscription. subscriptionPurchase in the project named my-fortune500-company-project. You have organized all your training code, from preprocessing data from the BigQuery table up to deploying the validated model to the Vertex AI endpoint, into a TensorFlow Extended (TFX) pipeline. You want to prevent prediction drift, i.e., a situation when a feature data distribution in production changes significantly over time. What should you do?
You recently developed a deep learning model using Keras, and now you are experimenting with different training
strategies. First, you trained the model using a single GPU, but the training process was too slow. Next, you
distributed the training across 4 GPUs using tf.distribute.MirroredStrategy (with no other changes), but you did not
observe a decrease in training time. What should you do?
You recently developed a deep learning model using Keras, and now you are experimenting with different training strategies. First, you trained the model using a single GPU, but the training process was too slow. Next, you distributed the training across 4 GPUs using tf.distribute.MirroredStrategy (with no other changes), but you did not observe a decrease in training time. What should you do?
You work for a gaming company that has millions of customers around the world. All games offer a chat feature that
allows players to communicate with each other in real time. Messages can be typed in more than 20 languages and
are translated in real time using the Cloud Translation API. You have been asked to build an ML system to moderate
the chat in real time while assuring that the performance is uniform across the various languages and without
changing the serving infrastructure.
You trained your first model using an in-house word2vec model for embedding the chat messages translated by the
Cloud Translation API. However, the model has significant differences in performance across the different
languages. How should you improve it?
You work for a gaming company that has millions of customers around the world. All games offer a chat feature that allows players to communicate with each other in real time. Messages can be typed in more than 20 languages and are translated in real time using the Cloud Translation API. You have been asked to build an ML system to moderate the chat in real time while assuring that the performance is uniform across the various languages and without changing the serving infrastructure. You trained your first model using an in-house word2vec model for embedding the chat messages translated by the Cloud Translation API. However, the model has significant differences in performance across the different languages. How should you improve it?
You work for a gaming company that develops massively multiplayer online (MM O) games. You built a TensorFlow model that predicts whether players will make in-app purchases of more than $10 in the next two weeks. The model’s predictions will be used to adapt each user’s game experience. User data is stored in BigQuery. How should you serve your model while optimizing cost, user experience, and ease of management?
You work for a gaming company that develops massively multiplayer online (MM O) games. You built a TensorFlow model that predicts whether players will make in-app purchases of more than $10 in the next two weeks. The model’s predictions will be used to adapt each user’s game experience. User data is stored in BigQuery. How should you serve your model while optimizing cost, user experience, and ease of management?
You are building a linear regression model on BigQuery ML to predict a customer’s likelihood of purchasing your
company’s products. Your model uses a city name variable as a key predictive component. In order to train and
serve the model, your data must be organized in columns. You want to prepare your data using the least amount of
coding while maintaining the predictable variables. What should you do?
You are building a linear regression model on BigQuery ML to predict a customer’s likelihood of purchasing your company’s products. Your model uses a city name variable as a key predictive component. In order to train and serve the model, your data must be organized in columns. You want to prepare your data using the least amount of coding while maintaining the predictable variables. What should you do?
You are an ML engineer at a bank that has a mobile application. Management has asked you to build an ML-based
biometric authentication for the app that verifies a customer’s identity based on their fingerprint. Fingerprints are
considered highly sensitive personal information and cannot be downloaded and stored into the bank databases.
Which learning strategy should you recommend to train and deploy this ML mode?
You are an ML engineer at a bank that has a mobile application. Management has asked you to build an ML-based biometric authentication for the app that verifies a customer’s identity based on their fingerprint. Fingerprints are considered highly sensitive personal information and cannot be downloaded and stored into the bank databases. Which learning strategy should you recommend to train and deploy this ML mode?
Flashcards
Capital of France (example flashcard)
Capital of France (example flashcard)
Paris