A) The @data is labeled, meaning each example is paired with a target output. B) The data is unlabeled, and the model must find patterns on its own. C) The data is generated randomly by the algorithm. D) The data is generated randomly by the algorithm.
A) Discover hidden patterns without any guidance. B) Gen@eralize from the training data to make accurate predictions on new, unseen data. C) Memorize the entire training dataset perfectly. D) Reduce the dimensionality of the input data for visualization.
A) The loss function. B) The@ label or target output. C) The input features. D) The model's parameters.
A) Diagnosing @a tumor as malignant or benign based on medical images. B) Estimating the annual revenue of a company. C) Forecasting the temperature for tomorrow. D) Predicting the selling price of a house based on its features.
A) Clustering problem. B) Regressio@n problem. C) Classification problem. D) Dimensionality reduction problem.
A) To achieve perfect accuracy on a held-out test set. B) To @discover the inherent structure, patterns, or relationships within unlabeled data. C) To predict a target variable based on labeled examples. D) To classify emails into spam and non-spam folders.
A) C@lustering. B) Classification. C) Regression. D) Reinforcement Learning.
A) Logistic Regression, a type of supervised learning. B) A support vector machine for classification. C) Linear Regression, a type of supervised learning. D) Clu@stering, a type of unsupervised learning.
A) Re@duce the number of features while preserving the most important information in the data. B) Increase the number of features to improve model accuracy. C) Predict a continuous output variable. D) Assign categorical labels to each data point.
A) Classification in supervised learning. B) Regression in supervised learning. C) Ass@ociation rule learning in unsupervised learning. D) Deep learning with neural networks.
A) It is always more accurate than fully supervised learning. B) It requires no labeled data at all. C) La@beling data is often expensive and time-consuming, so it leverages a small labeled set with a large unlabeled set. D) It is simpler to implement than unsupervised learning.
A) "Is this pattern anomalous?" B) "What is the underlying group?" C) "Which category?" D) "H@ow much?" or "How many?"
A) "Whic@h category?" or "What class?" B) "How much?" or "How many?" C) "What is the correlation between these variables?" D) "How can I reduce the number of features?"
A) Logistic Regression B) Decision Tree for classification C) Lin@ear Regression D) k-Nearest Neighbors for classification
A) Clustering B) Mult@i-class classification C) Regression D) Dimensionality reduction
A) Th@e final class labels or decisions B) The average value of a continuous target C) The input features for a new data point D) The probability of moving to the next node
A) A random number B) The name of the feature used for splitting C) A c@ontinuous value, often the mean of the target values of the training instances that reach the leaf D) A categorical class label
A) Guarantee to find the global optimum for any dataset B) Inter@pretability; the model's decision-making process is easy to understand and visualize C) Superior performance on all types of data compared to other algorithms D) Immunity to overfitting on noisy datasets
A) Perform linear regression more efficiently B) Grow a tree structure by making sequential decisions C) Fin@d a linear separating hyperplane in a high-dimensional feature space, even when the data is not linearly separable in the original space D) Initialize the weights of a neural network
A) The weights of a neural network layer B) . Da@ta points that are closest to the decision boundary and most critical for defining the optimal hyperplane C) The axes of the original feature space D) All data points in the training set
A) Their lower computational cost for very large datasets B) Their inherent resistance to any form of overfitting C) Their superior interpretability and simplicity D) The@ir effectiveness in high-dimensional spaces and their ability to model complex, non-linear decision boundaries
A) Clustering B) Data preprocessing C) Dimensionality reduction D) Tr@aining or model fitting
A) The algorithms are not well-defined B) There@ are no ground truth labels to compare the results against C) The data is always too small D) The models are always less accurate than supervised models
A) A Regression algorithm like Linear Regression B) Dimen@sionality Reduction techniques like Principal Component Analysis (PCA) C) An Association rule learning algorithm D) A Classification algorithm like Logistic Regression
A) Clus@tering, an unsupervised learning method B) A neural network for image recognition C) Regression, a supervised learning method D) Classification, a supervised learning method
A) Support vector B) Decision node in a tree C) Artifi@cial neuron or perceptron, which receives inputs, applies a transformation, and produces an output D) Principal component
A) Loss function B) Activ@ation function C) Optimization algorithm D) Kernel function
A) Rectifie@d Linear Unit (ReLU) B) The identity function (f(x) = x) C) A constant function D) The mean squared error function
A) Manually setting the weights based on expert knowledge B) Clustering the input data C) Randomly assigning weights and never changing them D) Iterativ@ely adjusting the weights and biases to minimize a loss function
A) Visualize the network's architecture B) Perform clustering on the output layer C) Initialize the weights before training D) Efficient@ly calculate the gradient of the loss function with respect to all the weights in the network, enabling the use of gradient descent
A) Neural n@etworks with many layers (hence "deep") B) K-means clustering exclusively C) Simple linear regression models D) Decision trees with a single split
A) Always train faster and with less data B) Operate without any need for data preprocessing C) Be perfectly interpretable, like a decision tree D) Auto@matically learn hi@erarchical feature representations from data
A) Text data and natural language processing B) Unsupervised clustering of audio signals C) Image @data, due to their architecture which exploits spatial locality D) Tabular data with many categorical features
A) Detect@ local features (like edges or textures) in the input by applying a set of learnable filters B) Initialize the weights of the network C) Perform the final classification D) Flatten the input into a single vector
A) Only image data B) Static, non-temporal data C) Independent and identically distributed (IID) data points D) Sequ@ential data, like time series or text, due to their internal "memory" of previous inputs
A) The@ gradients becoming exceedingly small as they are backpropagated through many layers, which can halt learning in early layers B) The gradients becoming too large and causing numerical instability C) The loss function reaching a perfect value of zero D) The model overfitting to the training data
A) Fit th@e model's parameters (e.g., the weights in a neural network) B) Deploy the model in a production environment C) Tune the model's hyperparameters D) Provide an unbiased evaluation of a final model's performance
A) Tun@ing hyperparameters and making decisions about the model architecture during development B) The initial training of the model's weights C) The final, unbiased assessment of the model's generalization error D) Data preprocessing and cleaning
A) Used repeatedly to tune the model's hyperparameters B) Ignored in the machine learning pipeline C) Use@d only once, for a final evaluation of the model's performance on unseen data after model development is complete D) Used repeatedly to tune the model's hyperparameters
A) Learns @the training data too well, including its noise and outliers, and performs poorly on new, unseen data B) Is evaluated using the training set instead of a test set C) Is too simple to capture the trends in the data D) Fails to learn the underlying pattern in the training data
A) Dropo@ut, which randomly ignores a subset of neurons during training B) Increasing the model's capacity by adding more layers C) Using a smaller training dataset D) Training for more epochs without any checks
A) The activation function used in the output layer B) The weights connecting the input layer to the hidden layer C) The error from sensitivity to small fluctuations in the training set, leading to overfitting D) The erro@r from erroneous assumptions in the learning algorithm, leading to underfitting
A) The er@ror from sensitivity to small fluctuations in the training set, leading to overfitting B) The intercept term in a linear regression model C) The speed at which the model trains D) The error from erroneous assumptions in the learning algorithm, leading to underfitting
A) Bias and variance can be minimized to zero simultaneously B) Only variance is important for model performance C) Decrea@sing bias will typically increase variance, and vice versa. The goal is to find a balance D) Only bias is important for model performance
A) A well-generalized model B) Underfitting C) Perfect model performance D) Overf@itting
A) The number of layers in the network B) The speed of the backpropagation algorithm C) How well the model is performing on the training data; it's the quantity we want to minimize during training D) The accuracy on the test set
A) Guarantees finding the global minimum for any loss function B) Is only used for unsupervised learning C) Iteratively adjusts parameters in the direction that reduces the loss function D) Randomly searches the parameter space for a good solution
A) The number of layers in a neural network B) The amount of training data used in each epoch C) The size of the step taken during each parameter update. A rate that is too high can cause divergence, while one that is too low can make training slow D) The activation function for the output layer
A) A type of regularization technique B) The processing of a single training example C) The final evaluation on the test set D) One complete pass of the entire training dataset through the learning algorithm
A) The number of training examples used in one forward/backward pass before the model's parameters are updated B) The total number of examples in the training set C) The number of validation examples D) The number of layers in the network
A) One complete pass of the entire training dataset through the learning algorithm B) The processing of a single training example
A) One complete pass of the entire training dataset through the learning algorithm B) The processing of a single training example
A) One complete pass of the entire training dataset through the learning algorithm B) One complete pass of the entire training dataset through the learning algorithm
A) One complete pass of the entire training dataset through the learning algorithm B) One complete pass of the entire training dataset through the learning algorithm
A) The processing of a single training example B) One complete pass of the entire training dataset through the learning algorithm
A) The processing of a single training example B) One complete pass of the entire training dataset through the learning algorithm
A) The processing of a single training example B) One complete pass of the entire training dataset through the learning algorithm
A) One complete pass of the entire training dataset through the learning algorithm B) The processing of a single training example
A) The processing of a single training example B) One complete pass of the entire training dataset through the learning algorithm
A) One complete pass of the entire training dataset through the learning algorithm B) The processing of a single training example
A) One complete pass of the entire training dataset through the learning algorithm B) The processing of a single training example
A) The processing of a single training example B) One complete pass of the entire training dataset through the learning algorithm
A) The processing of a single training example B) One complete pass of the entire training dataset through the learning algorithm
A) The processing of a single training example B) One complete pass of the entire training dataset through the learning algorithm
A) The processing of a single training example B) One complete pass of the entire training dataset through the learning algorithm |