The AWS Certified Machine Learning - Specialty certification is intended for individuals who perform a development or data science role. It validates a candidate's ability to design, implement, deploy, and maintain machine learning (ML) solutions for given business problems.
Use AWS Data Pipeline to transform the data and Amazon RDS to run queries.
Use AWS Glue to catalogue the data and Amazon Athena to run queries.
Use AWS Batch to run ETL on the data and Amazon Aurora to run the queries.
Use AWS Lambda to transform the data and Amazon Kinesis Data Analytics to run queries.
Rate this question:
Use AWS Data Pipeline to transform the data and Amazon RDS to run queries.
Use AWS Glue to catalogue the data and Amazon Athena to run queries.
Use AWS Batch to run ETL on the data and Amazon Aurora to run the queries.
Use AWS Lambda to transform the data and Amazon Kinesis Data Analytics to run queries.
Rate this question:
Use the Synthetic Minority Oversampling Technique (SMOTE) to oversample the fraud records
Undersample the non-fraudulent records to improve the class imbalance
Use K-fold cross validation when training the model
Drop all the fraud examples, and use a One-Class SVM to classify
Rate this question:
Increase the instance size for training
Increase the batch size in the model
Change the input mode to Pipe
Create an Amazon EBS volume with the data on it and attach it to the training job
Rate this question:
Binarization
One-hot encoding
Tokenization
Normalization transformation
Rate this question:
Early stopping
Random initialization of weights with appropriate seed
Increasing the number of epochs
Adding another layer with the 100 neurons
Rate this question:
A custom CNN model
An LSTM model
Amazon Textract
Amazon Personalize
Rate this question:
Amazon Transcribe, Amazon Translate, and Amazon Comprehend
Amazon Transcribe, Amazon Comprehend, and Amazon SageMaker seq2seq
Amazon Transcribe, Amazon Translate, and Amazon SageMaker Neural Topic Model (NTM)
Amazon Transcribe, Amazon Translate, and Amazon SageMaker BlazingText
Rate this question:
Increase the training data by adding variation in rotation for training images.
Increase the number of epochs for model training.
Increase the number of layers for the neural network.
Increase the dropout rate for the second-to-last layer.
Rate this question:
Amazon S3 and Amazon Athena
Amazon Redshift and AWS Glue
Amazon DynamoDB and DynamoDB Accelerator (DAX)
Amazon RDS and Amazon ES
Rate this question:
AWS CloudTrail
AWS Health
AWS Trusted Advisor
Amazon CloudWatch
AWS Config
Rate this question:
Oversampling using bootstrapping
Undersampling
Oversampling using SMOTE
Class weight adjustment
Rate this question:
Amazon CloudWatch
Amazon SageMaker
Amazon EMR with Spark
Amazon Kinesis Data Analytics
Rate this question:
The model needs to be completely re-engineered because it is unable to handle product inventory changes
The model's hyperparameters should be periodically updated to prevent drift
The model should be periodically retrained from scratch using the original data while adding a regularization term to handle product inventory changes
The model should be periodically retrained using the original training data plus new data as product inventory changes
Rate this question:
Dropout
Smooth L1 loss
Softmax
Rectified linear units (ReLU)
Rate this question:
Build a content-based filtering recommendation engine with Apache Spark ML on Amazon EMR.
Build a collaborative filtering recommendation engine with Apache Spark ML on Amazon EMR.
Build a model-based filtering recommendation engine with Apache Spark ML on Amazon EMR.
Build a combinative filtering recommendation engine with Apache Spark ML on Amazon EMR.
Rate this question:
Latent Dirichlet Allocation (LDA)
Recurrent neural network (RNN)
K-means
Convolutional neural network (CNN)
Rate this question:
Recall
Misclassification rate
Mean absolute percentage error (MAPE)
Area Under the ROC Curve (AUC)
Rate this question:
Regression
Classification
Recommender system
Reinforcement learning
Rate this question:
Yes, because it is generalizing well on the training set
No, because it is generalizing well on the training set
No, because it is not generalizing well on the test set
Yes, because it is not generalizing well on the test set
Rate this question:
Perform one-hot encoding on highly correlated features
Use matrix multiplication on highly correlated features.
Create a new feature space using principal component analysis (PCA)
Apply the Pearson correlation coefficient
Rate this question:
Latent Dirichlet Allocation (LDA) for the given collection of discrete data to identify patterns in the customer database.
A neural network with a minimum of three layers and random initial weights to identify patterns in the customer database
Collaborative filtering based on user interactions and correlations to identify patterns in the customer database
Random Cut Forest (RCF) over random subsamples to identify patterns in the customer database
Rate this question:
Apply quantile binning to group the data into categorical bins to keep any relationships in the data by replacing the magnitude with distribution.
Apply the Cartesian product transformation to create new combinations of fields that are independent of the magnitude.
Apply normalization to ensure each field will have a mean of 0 and a variance of 1 to remove any significant magnitude.
Apply the orthogonal sparse Diagram (OSB) transformation to apply a fixed-size sliding window to generate new features of a similar magnitude.
Rate this question:
Long short-term memory (LSTM) model with scaled exponential linear unit (SELL))
Logistic regression
Support vector machine (SVM) with non-linear kernel
Single perceptron with tanh activation function
Rate this question:
SageMaker Random Cut Forest
Kinesis Data Streams Naive Bayes Classifier
Kinesis Data Analytics Random Cut Forest
Kinesis Data Analytics Nearest Neighbor
Rate this question:
Review SageMaker logs that have been written to Amazon S3 by leveraging Amazon Athena and Amazon OuickSight to visualize logs as they are being produced.
Generate an Amazon CloudWatch dashboard to create a single view for the latency, memory utilization, and CPU utilization metrics that are outputted by Amazon SageMaker.
Build custom Amazon CloudWatch Logs and then leverage Amazon ES and Kibana to query and visualize the data as it is generated by Amazon SageMaker.
Send Amazon CloudWatch Logs that were generated by Amazon SageMaker lo Amazon ES and use Kibana to query and visualize the log data.
Rate this question:
.Store datasets as files in Amazon S3.
Store datasets as files in an Amazon EBS volume attached to an Amazon EC2 instance.
Store datasets as tables in a multi-node Amazon Redshift cluster.
Store datasets as global tables in Amazon DynamoDB.
Rate this question:
Listwise deletion
Last observation carried forward
Multiple imputations
Mean substitution
Rate this question:
Increase the randomization of training data in the mini-batches used in training.
Allocate a higher proportion of the overall data to the training dataset
Apply L1 or L2 regularization and dropouts to the training.
Reduce the number of layers and units (or neurons) from the deep learning network.
Rate this question:
An Amazon EBS-backed Amazon EC2 instance with hourly directories
An Amazon RDS database with hourly table partitions
An Amazon S3 data lake with hourly object prefixes
An Amazon EMR cluster with hourly hive partitions on Amazon EBS volumes
Rate this question:
Use Amazon Comprehend with the transcribed files to build the key topics.
Use Amazon Translate with the transcribed files to train and build a model for the key topics.
Use the AWS Deep Learning AMI with Gluon Semantic Segmentation on the transcribed files to train and build a model for the key topics.
Use the Amazon SageMaker k-Nearest-Neighbors (kNN) algorithm on the transcribed files to generate a word embeddings dictionary for the key topics.
Rate this question:
The model predicts both the trend and the seasonality well.
The model predicts the trend well, but not the seasonality.
The model predicts the seasonality well, but not the trend.
The model does not predict the trend or the seasonality well.
Rate this question:
K-means clustering
Random Cut Forest (RCF)
XGBoost
BlazingText
Rate this question:
Uncheck the option for Internet access when creating your notebook instance, and it will handle the rest automatically.
No action is required, the VPC will block the notebook instances from accessing the Internet.
Use IAM to restrict Internet access from the notebook instance.
Disable direct Internet access when specifying the VPC for your notebook instance, and use VPC interface endpoints (PrivateLink) to allow the connections needed to train and host your model. Modify your instance's security group to allow outbound connections for training and hosting.
Rate this question:
The class distribution in the dataset is imbalanced
Dataset shuffling is disabled
The batch size is too big
The learning rate is very high
Rate this question:
Ingest the data using Amazon Kinesis Data Firehose, and use Amazon Kinesis Data Analytics Random Cut Forest (RCF) for anomaly detection. Then use Kinesis Data Firehose to stream the results to Amazon S3.
Ingest the data into Apache Spark Streaming using Amazon EMR. and use Spark MLlib with k-means to perform anomaly detection. Then store the results in an Apache Hadoop Distributed File System (HDFS) using Amazon EMR with a replication factor of three as the data lake.
Ingest the data and store it in Amazon S3 Use AWS Batch along with the AWS Deep Learning AMIs to train a k-means model using TensorFlow on the data in Amazon S3.
Ingest the data and store it in Amazon S3. Have an AWS Glue job that is triggered on demand transform the new data. Then use the built-in Random Cut Forest (RCF) model within Amazon SageMaker to detect anomalies in the data.
Rate this question:
XGBoost is an Extreme Gradient Boosting algorithm that is optimized for boosted decision trees
XGBoost is a state-of-the-art algorithm that uses logistic regression to split each feature of the data basedon certain conditions
XGBoost is a robust, flexible, scalable algorithm that uses logistic regression to classify data into buckets
XGBoost is an efficient and scalable neural network architecture.
Rate this question:
Cross-entropy log loss
Sigmoid
Root Mean squared error (RMSE)
Precision
Rate this question:
Create a data lake using Amazon S3 as the data storage layer
Store unstructured data in Amazon DynamoDB and structured data in Amazon RDS
Use Amazon FSx to host the data for training
Use Amazon Elastic Block Store (Amazon EBS) volumes to store the data with data backup
Rate this question:
Write a direct connection to the SQL database within the notebook and pull data in.
Push the data from Microsoft SQL Server to Amazon S3 using an AWS Data Pipeline and provide the S3 location within the notebook.
Move the data to Amazon DynamoDB and set up a connection to DynamoDB within the notebook to pull data in.
Move the data to Amazon ElastiCache using AWS DMS and set up a connection within the notebook to pull data in for fast access.
Rate this question:
Logistic regression
Random Cut Forest (RCF)
Principal component analysis (PCA)
Linear regression
Rate this question:
Receiver operating characteristic (ROC) curve
Misclassification rate
Root Mean Square Error (RM&)
L1 norm
Rate this question:
Use IAM roles. “logs:*” are added to those IAM roles.
Enable AWS CloudTrail.
Enable CloudWatch logs.
Use the Amazon SageMaker SDK to call the ‘sagemaker_history()’ function.
Rate this question:
As variable 1 increases, variable 5 increases
As variable 1 increases, variable 5 decreases
Variable 1 does not have any influence on variable 5
The data is not sufficient to make a well-informed interpretation
Rate this question:
Regression
Classification
Recommender system
Reinforcement learning
Rate this question:
Use a custom encryption algorithm to encrypt the data and store the data on an Amazon SageMaker instance in a VPC. Use the SageMaker DeepAR algorithm to randomize the credit card numbers.
Use an IAM policy to encrypt the data on the Amazon S3 bucket and Amazon Kinesis to automatically discard credit card numbers and insert fake credit card numbers.
Use an Amazon SageMaker launch configuration to encrypt the data once it is copied to the SageMaker instance in a VPC. Use the SageMaker principal component analysis (PCA) algorithm to reduce the length of the credit card numbers.
Use AWS KMS to encrypt the data on Amazon S3 and Amazon SageMaker, and redact the credit card numbers from the customer data with AWS Glue.
Rate this question:
1
4
7
10
Rate this question:
A longer training time
Making the network larger
Using a different optimizer
Using some form of regularization
Rate this question:
Quiz Review Timeline (Updated): Jan 8, 2025 +
Our quizzes are rigorously reviewed, monitored and continuously updated by our expert board to maintain accuracy, relevance, and timeliness.
Wait!
Here's an interesting quiz for you.