1.
multiclass classification diagram
2.
linear regression and scatter plots
3.
pivot table
4.
K-means cluster diagram
Q 1 / 92
1.
The product would look for new patterns in spam messages.
2.
The product could go through the keyword list much more quickly.
3.
The product could have a much longer keyword list.
4.
The product could find spam messages using far fewer keywords.
Q 2 / 92
1.
data cluster
2.
Supervised set
3.
big data
4.
test data
Q 3 / 92
1.
patterns
2.
programs
3.
rules
4.
data
Q 4 / 92
1.
It was consistently wrong.
2.
It was inconsistently wrong.
3.
It was consistently right.
4.
It was equally right end wrong.
Q 5 / 92
1.
Find labeled data of sunny days so that the machine will learn to identify bad weather.
2.
Use unsupervised learning have the machine look for anomalies in a massive weather database.
3.
Create a training set of unusual patterns and ask the machine learning algorithms to classify them.
4.
Create a training set of normal weather and have the machine look for similar patterns.
Q 6 / 92
1.
regression
2.
boosting
3.
bagging
4.
stacking
Q 7 / 92
1.
Regression analysis
2.
K-means clustering
3.
Big data
4.
Unsupervised learning
Q 8 / 92
1.
a data entry system
2.
a data warehouse system
3.
a massive data repository
4.
a product recommendation system
Q 9 / 92
![Machine Learning Q10](images/machine-learning_Q10.jpg)
1.
a decision tree
2.
reinforcement learning
3.
K-nearest neighbor
4.
a clear trendline
Q 10 / 92
1.
The algorithms would help the meters access the internet.
2.
The algorithms will improve the wireless connectivity.
3.
The algorithms would help your organization see patterns of the data.
4.
By using machine learning algorithms, you are creating an IoT device.
Q 11 / 92
1.
regression
2.
clustering
3.
classification
4.
dimensionality reduction
Q 12 / 92
1.
It naively assumes that you will have no data.
2.
It does not even try to create accurate predictions.
3.
It naively assumes that the predictors are independent from one another.
4.
It naively assumes that all the predictors depend on one another.
Q 13 / 92
![Machine Learning Q14](images/machine-learning_Q14.jpg)
1.
It is a linear regression chart.
2.
It is a supervised trendline chart.
3.
It is a decision tree.
4.
It is a clustering trend chart.
Q 14 / 92
1.
Artificial intelligence focuses on classification, while machine learning is about clustering data.
2.
Machine learning is a type of artificial intelligence that relies on learning through data.
3.
Artificial intelligence is form of unsupervised machine learning.
4.
Machine learning and artificial intelligence are the same thing.
Q 15 / 92
1.
The algorithms are typically run more powerful servers.
2.
The algorithms are better at seeing patterns in the data.
3.
Machine learning servers can host larger databases.
4.
The algorithms can run on unstructured data.
Q 16 / 92
1.
Create an artificial neural network that would host the company directory.
2.
Use machine learning to better predict risk.
3.
Create an algorithm that consolidates all of your Excel spreadsheets into one data lake.
4.
Use machine learning and big data to research salary requirements.
Q 17 / 92
![Machine Learning Q18](images/machine-learning_Q18.jpg)
1.
Training Set
2.
Unsupervised Data
3.
Supervised Learning
4.
Binary Classification
Q 18 / 92
1.
You will almost certainly underfit the model.
2.
You will pick the wrong algorithm.
3.
You might not have enough data for both.
4.
You will almost certainly overfit the model.
Q 19 / 92
**Explanation**: While machine learning algorithms don't have bias, the data can have them.
1.
Machine learning algorithms are based on math and statistics, and so by definition will be unbiased.
2.
There is no way to identify bias in the data.
3.
Machine learning algorithms are powerful enough to eliminate bias from the data.
4.
All human-created data is biased, and data scientists need to account for that.
Q 20 / 92
1.
The predictions of one model become the inputs another.
2.
You use different versions of machine learning algorithms.
3.
You use several machine learning algorithms to boost your results.
4.
You stack your training set and testing set together.
Q 21 / 92
1.
training data
2.
linear regression
3.
big data
4.
test data
Q 22 / 92
**Explanation**: The problem explicitly states "clustering".
1.
centroid reinforcement
2.
K-nearest neighbor
3.
binary classification
4.
K-means clustering
Q 23 / 92
1.
Include training email data from all employees.
2.
Include training email data from new employees.
3.
Include training email data from seasoned employees.
4.
Include training email data from employees who write the majority of internal emails.
Q 24 / 92
1.
unsupervised machine learning
2.
binary classification
3.
supervised machine learning
4.
reinforcement learning
Q 25 / 92
![Machine Learning Q26](images/machine-learning_Q26.jpg) Note: there are centres of clusters (C0, C1, C2).
1.
K-nearest neighbor
2.
a decision tree
3.
a linear regression
4.
a K-means cluster
Q 26 / 92
1.
aggregated trees
2.
boosted trees
3.
bagged trees
4.
stacked trees
Q 27 / 92
1.
semi-supervised learning
2.
supervised learning
3.
reinforcement learning
4.
unsupervised learning
Q 28 / 92
1.
In K-means clustering, the initial centroids are sometimes randomly selected.
2.
K-means clustering is often used in supervised machine learning.
3.
The number of clusters are always randomly selected.
4.
To be accurate, you want your centroids outside of the cluster.
Q 29 / 92
1.
supervised learning
2.
semi-supervised learning
3.
reinforcement learning
4.
unsupervised learning
Q 30 / 92
1.
random forest
2.
logistic regression
3.
KNN
4.
deep neural network
Q 31 / 92
1.
Higher K values will produce noisy data.
2.
Higher K values lower the bias but increase the variance.
3.
Higher K values need a larger training set.
4.
Higher K values lower the variance but increase the bias.
Q 32 / 92
1.
supervised learning
2.
unsupervised learning
3.
reinforcement learning
4.
semi-unsupervised learning
Q 33 / 92
1.
It uses unsupervised learning to cluster together transactions and unsupervised learning to classify the customers.
2.
It uses only unsupervised machine learning.
3.
It uses supervised learning to create clusters and unsupervised learning for classification.
4.
It uses reinforcement learning to classify the customers.
Q 34 / 92
1.
high variance and low bias
2.
low bias and low variance
3.
low variance and high bias
4.
high bias and high variance
Q 35 / 92
1.
No, data model bias and variance are only a challenge with reinforcement learning.
2.
Yes, data model bias is a challenge when the machine creates clusters.
3.
Yes, data model variance trains the unsupervised machine learning algorithm.
4.
No, data model bias and variance involve supervised learning.
Q 36 / 92
**Explanation:** Logistic regression is far better than linear regression at binary classification since it biases the result toward one extreme or the other. K-means clustering can be used for classification but is not as accurate in most scenarios.
1.
K-means
2.
Logistic regression
3.
Linear regression
4.
Principal Component Analysis (PCA)
Q 37 / 92
**Explanation**: This one is pretty straight forward and a fundamental concept.
1.
supervised learning
2.
data
3.
unsupervised learning
4.
algorithms
Q 38 / 92
1.
It will take too long for programmers to scrub poor data.
2.
If the data is high quality, the algorithms will be easier to develop.
3.
Low-quality data requires much more processing power than high-quality data.
4.
If the data is low quality, you will get inaccurate results.
Q 39 / 92
1.
share common characteristics
2.
be part of the root node
3.
have a Euclidean connection
4.
be part of the same cluster
Q 40 / 92
1.
Reinforcement machine learning
2.
unsupervised machine learning
3.
supervised machine learning
4.
semi-supervised machine learning
Q 41 / 92
**Explanation**: This one is similar to an example talked about in the Stanford Machine Learning course.
1.
You will be able to prioritize different classes of drugs, such as antibiotics.
2.
You can create a training set of drugs you would like to discover.
3.
The algorithms will cluster together drugs that have similar traits.
4.
Human experts can create classes of drugs to help guide discovery.
Q 42 / 92
1.
The system went from supervised learning to reinforcement learning.
2.
The system evolved from supervised learning to unsupervised learning.
3.
The system evolved from unsupervised learnin9 to supervised learning.
4.
The system evolved from reinforcement learning to unsupervised learning.
Q 43 / 92
1.
It could better protect against undiscovered threats.
2.
It would very likely lower the hardware requirements.
3.
It would substantially shorten your development time.
4.
It would increase the speed of the appliance.
Q 44 / 92
1.
Use reinforcement learning to reward the system when a new person participates.
2.
Use unsupervised machine learning to cluster together people based on patterns the machine discovers.
3.
Use Supervised machine learning to sort people by demographic data.
4.
Use Supervised machine learning to classify people by body temperature.
Q 45 / 92
1.
statistics
2.
structured data
3.
availability
4.
algorithms
Q 46 / 92
![Machine Learning Q45](images/machine-learning_Q45.jpg)
1.
unsupervised learning
2.
complex cluster
3.
multiclass classification
4.
k-nearest neighbour
Q 47 / 92
1.
deep learning artificial neural network that relies on petabytes of data
2.
unsupervised machine learning system that clusters together the best candidates
3.
Not recommend machine learning for this project
4.
supervised machine learning system that classifies applicants into existing groups // we do not need to classify best candidates we just need to classify job applicants in to existing categories
Q 48 / 92
1.
regression analysis
2.
unsupervised learning
3.
high -variance modeling
4.
ensemble modeling
Q 49 / 92
1.
machine learning algorithm
2.
training set
3.
big data test set
4.
data cluster
Q 50 / 92
1.
You are overfitting the model to the data
2.
You need a smaller training set
3.
You are underfitting the model to the data
4.
You need a larger training set
Q 51 / 92
1.
an unsupervised machine learning system that clusters together the best candidates.
2.
you would not recommend a machine learning system for this type of project.
3.
a deep learning artificial neural network that relies on petabytes of employment data.
4.
a supervised machine learning system that classifies applicants into existing groups.
Q 52 / 92
1.
you use it as your training set.
2.
You label it big data.
3.
You split it into a training set and test set.
4.
You use it as your test set.
Q 53 / 92
1.
semi-supervised machine learning
2.
supervised machine learning
3.
unsupervised machine learning
4.
reinforcement learning
Q 54 / 92
1.
Batch learning
2.
Offline learning
3.
Both A and B
4.
None of the above
Q 55 / 92
1.
Decision Tree
2.
Linear Regression
3.
PCA
4.
Naive Bayesian
Q 56 / 92
1.
Decision Trees
2.
K-means clustering
3.
Density-based clustering
4.
Model-based clustering
Q 57 / 92
1.
The entropy function.
2.
The squared error.
3.
The cross-entropy function.
4.
The number of mistakes.
Q 58 / 92
1.
Higher
2.
same
3.
Lower
4.
it could be any of the above
Q 59 / 92
1.
good fitting
2.
overfitting
3.
underfitting
4.
all of the above
Q 60 / 92
![Machine Learning Q58](images/machine-learning_Q58.jpg) **Explanation**: Shows data being classified into more than two categories or classes. Thus, this is a multi-class classification challenge.
1.
This is a multiclass classification challenge.
2.
This is a multi-binary classification challenge.
3.
This is a binary classification challenge.
4.
This is a reinforcement classification challenge.
Q 61 / 92
`Underfitted data models usually have high bias and low variance. Overfitted data models have low bias and high variance.`
1.
There is too little data in your training set.
2.
There is too much data in your training set.
3.
There is not a lot of variance but there is a high bias.
4.
Your model has low bias but high variance.
Q 62 / 92
`The answer is self-explanatory: if Asian users are the only group of people making the complaint, then the training data should have more Asian faces.`
1.
Include Asian faces in your test data and retrain your model.
2.
Retrain your model with updated hyperparameter values.
3.
Retrain your model with smaller batch sizes.
4.
Include Asian faces in your training data and retrain your model.
Q 63 / 92
**Explanation**: // This question is very similar to Q49 but involves a polar opposite scenario. `that answer somewhat vague and unsettled. Small number of matchings does not necessarily implies that the model overfits, especially given 500 (!) independent variables. To me, it sounds more reasonable that the threshold (matching) criterion might be too tight, thus allowing only a small number of matching to occur. So a solution can be either softening the threshold criterion or increasing the number of candidates.`
1.
Your training set is too large.
2.
You are underfitting the model to the data.
3.
You are overfitting the model to the data.
4.
Your machine is creating inaccurate clusters.
Q 64 / 92
1.
What kernels extract
2.
Feature Maps
3.
How kernels Look
Q 65 / 92
![image](images/machine-learning_Q62.png)
1.
76%
2.
88%
3.
12%
4.
0.0008%
Q 66 / 92
1.
Wise fill-in of controlled random values
2.
Replace missing values with averaging across all samples
3.
Remove defective samples
4.
Imputation
Q 67 / 92
1.
SVM
2.
PCA
3.
LDA
4.
TSNE
Q 68 / 92
1.
Capturing complex non-linear patterns
2.
Transforming continuous values into "ON" (1) or "OFF" (0) values
3.
Help avoiding the vanishing/exploding gradient problem
4.
Their ability to activate each neurons individually.
Q 69 / 92
1.
kullback-leibler (KL) loss
2.
Binary Crossentropy
3.
Mean Squared Error (MSE)
4.
Any L2 loss
Q 70 / 92
![image](images/machine-learning_Q67.png) | no. | Red | Blue | Green | | --- | --- | --- | --- | | **1.** | Validation error | Training error | Test error | | **2.** | Training error | Test error | Validation error | | **3.** | Optimal error | Validation error | Test error | | **4.** | Validation error | Training error | Optimal error |
1.
1
2.
2
3.
3
4.
4
Q 71 / 92
`// these nodes decide whether the someone decides to go to beach or not, for example if its rainy people will mostly refrain from going to beach`
1.
tree nodes
2.
predictors
3.
root nodes
4.
deciders
Q 72 / 92
1.
Set up a cluster of machines to label the images
2.
Create a subset of the images and label then yourself
3.
Use naive Bayes to automatically generate labels.
4.
Hire people to manually label the images
Q 73 / 92
![image](images/machine-learning_Q70.png) `// since the data is accurately classified and is neither overfitting or underfitting the dataset`
1.
low bias, high variance
2.
high bias, low variance
3.
high bias, high variance
4.
low bias, low variance
Q 74 / 92
1.
structured data
2.
algorithms
3.
time
4.
computer scientists
Q 75 / 92
1.
Scikit-learn
2.
PyTorch
3.
Tensowflow Lite
4.
Tensorflow
Q 76 / 92
1.
a spreadsheet
2.
20,000 recorded voicemail messages
3.
100,000 images of automobiles
4.
hundreds of gigabytes of audio files
Q 77 / 92
1.
confidence
2.
alpha
3.
power
4.
significance
Q 78 / 92
1.
naive Bayes classifier
2.
K-nearest neighbor
3.
multiclass classification
4.
decision tree
Q 79 / 92
1.
when the machine learning algorithms do most of the programming
2.
when you don't do any data scrubbing
3.
when the learning happens continuously
4.
when you run your computation in one big instance at the beginning
Q 80 / 92
**Explanation**:Q-learning is a model-free reinforcement learning algorithm.Q-learning is a values-based learning algorithm. Value based algorithms updates the value function based on an equation(particularly Bellman equation).
1.
supervised machine learning with rewards
2.
a type of unsupervised learning that relies heavily on a well-established model
3.
a type of reinforcement learning where accuracy degrades over time
4.
a type of reinforcement learning that focuses on rewards
Q 81 / 92
1.
They would be grouped tightly together in the predicted outcome.
2.
They would be grouped tightly together but far from the predicted.
3.
They would be scattered around the predict outcome.
4.
They would be scattered far away from the predicted outcome.
Q 82 / 92
1.
unsupervised learning
2.
semi-supervised learning
3.
supervised learning
4.
semi-reinforcement learning
Q 83 / 92
1.
binary learning
2.
supervised learning
3.
unsupervised learning
4.
reinforcement learning
Q 84 / 92
`//You could use a naïve Bayes algorithm, to differentiate three classes of dog breeds — terrier, hound, and sport dogs. Each class has three predictors — hair length, height, and weight. The algorithm does something called class predictor probability.`
1.
multiclass binary classification
2.
naive Bayes
3.
unsupervised classification
4.
decision tree analysis
Q 85 / 92
1.
Create training clusters
2.
Remove predictors
3.
Use reinforcement learning
4.
Add more predictors
Q 86 / 92
1.
Suggest that the team is probably underfitting the model to the data.
2.
Suggest that unsupervised learning will lead to more interesting results.
3.
Make sure that they are picking the correct machine learning algorithms.
4.
Encourage the team to ask more interesting questions.
Q 87 / 92
1.
Unstructured data is always text.
2.
Unstructured data is much easier to store.
3.
Structured data has clearly defined data types.
4.
Sturctured data is much more popular.
Q 88 / 92
1.
Using an unsupervised machine learning algorithm to cluster together all the photographs.
2.
Crate a data lake with an unsupervised machine learning algorithm.
3.
Use a combination of unsupervised and supervised machine learning to create machine-defined data clusters.
4.
Use supervised machine learning to classify photographs based on a predetermined training set.
Q 89 / 92
1.
the under/over challenge
2.
balance between clustering classification
3.
bias-variance trade-off
4.
the multiclass training set challenge
Q 90 / 92
1.
the probability that doing one thing has an impact on another thing
2.
the probability that certain conditions are met
3.
the probability that, based on certain conditions, something will always be incorrect
4.
the probability of something being the correct answer
Q 91 / 92
1.
conditional
2.
multiclass
3.
independent
4.
binary
Q 92 / 92