MachineLearningAlgorithms / data /Ridge_Regression.json
deedrop1140's picture
Upload 41 files
0d00d62 verified
{
"title": "Ridge Regression Mastery: 100 MCQs",
"description": "A comprehensive set of multiple-choice questions designed to teach and test your understanding of Ridge Regression, starting from basic concepts to advanced scenario-based problems.",
"questions": [
{
"id": 1,
"questionText": "What is the main purpose of Ridge Regression?",
"options": [
"To reduce bias in predictions",
"To prevent overfitting by adding L2 regularization",
"To increase the complexity of the model",
"To reduce the number of features"
],
"correctAnswerIndex": 1,
"explanation": "Ridge Regression adds L2 regularization to penalize large coefficients, helping prevent overfitting."
},
{
"id": 2,
"questionText": "Which term is added to the loss function in Ridge Regression?",
"options": [
"Sum of squared residuals",
"Sum of absolute values of coefficients",
"Sum of squares of coefficients multiplied by alpha",
"Log-likelihood term"
],
"correctAnswerIndex": 2,
"explanation": "Ridge Regression adds alpha * sum of squared coefficients to the standard squared error loss."
},
{
"id": 3,
"questionText": "Ridge Regression is a type of:",
"options": [
"Linear Regression with L1 regularization",
"Linear Regression with L2 regularization",
"Logistic Regression",
"Decision Tree Regression"
],
"correctAnswerIndex": 1,
"explanation": "Ridge Regression is Linear Regression with L2 regularization to shrink coefficients."
},
{
"id": 4,
"questionText": "Which problem does Ridge Regression primarily address?",
"options": [
"Underfitting",
"Overfitting due to multicollinearity",
"Non-linear data",
"Categorical features"
],
"correctAnswerIndex": 1,
"explanation": "Ridge Regression reduces overfitting when features are highly correlated."
},
{
"id": 5,
"questionText": "How does Ridge Regression shrink coefficients?",
"options": [
"By adding noise to data",
"By adding a penalty proportional to the square of coefficients",
"By removing features randomly",
"By using stepwise regression"
],
"correctAnswerIndex": 1,
"explanation": "The L2 penalty in Ridge Regression discourages large coefficients."
},
{
"id": 6,
"questionText": "What happens if alpha=0 in Ridge Regression?",
"options": [
"It becomes standard Linear Regression",
"It becomes Lasso Regression",
"It ignores the bias term",
"It fails to converge"
],
"correctAnswerIndex": 0,
"explanation": "With alpha=0, the L2 penalty is removed, so Ridge Regression is equivalent to Linear Regression."
},
{
"id": 7,
"questionText": "Ridge Regression is particularly useful when:",
"options": [
"The dataset has multicollinearity among features",
"The dataset has very few samples",
"There is no noise in data",
"You want sparse coefficients"
],
"correctAnswerIndex": 0,
"explanation": "Ridge Regression handles multicollinearity by penalizing large correlated coefficients."
},
{
"id": 8,
"questionText": "Which metric is commonly used to select the optimal alpha in Ridge Regression?",
"options": [
"R-squared",
"Mean Squared Error on cross-validation",
"Correlation coefficient",
"Number of features selected"
],
"correctAnswerIndex": 1,
"explanation": "Cross-validation MSE is used to find the alpha that balances bias and variance."
},
{
"id": 9,
"questionText": "What effect does increasing the alpha parameter have?",
"options": [
"Increases overfitting",
"Decreases coefficient values and reduces overfitting",
"Increases model complexity",
"Removes features automatically"
],
"correctAnswerIndex": 1,
"explanation": "Higher alpha increases the penalty on large coefficients, which shrinks them and reduces overfitting."
},
{
"id": 10,
"questionText": "Why should features be standardized before applying Ridge Regression?",
"options": [
"To make computation faster",
"To give all features equal importance in regularization",
"To reduce number of samples",
"To convert all values to integers"
],
"correctAnswerIndex": 1,
"explanation": "Standardization ensures the penalty treats all features fairly, regardless of scale."
},
{
"id": 11,
"questionText": "Ridge Regression cannot produce sparse models because:",
"options": [
"It uses L1 penalty",
"It uses L2 penalty which shrinks but does not set coefficients to zero",
"It ignores regularization",
"It only works with one feature"
],
"correctAnswerIndex": 1,
"explanation": "L2 penalty reduces coefficient magnitudes but does not eliminate features completely."
},
{
"id": 12,
"questionText": "Which scenario favors Ridge Regression over Lasso?",
"options": [
"You want feature selection",
"All features are relevant and correlated",
"You have very few samples",
"Your target variable is categorical"
],
"correctAnswerIndex": 1,
"explanation": "Ridge is better when all features contribute and are correlated; Lasso performs feature selection."
},
{
"id": 13,
"questionText": "Which of the following is a loss function of Ridge Regression?",
"options": [
"Sum of squared errors",
"Sum of squared errors + alpha * sum of squared coefficients",
"Sum of absolute errors",
"Mean absolute percentage error"
],
"correctAnswerIndex": 1,
"explanation": "Ridge adds the L2 penalty to the usual squared error loss function."
},
{
"id": 14,
"questionText": "Scenario: Your data has 200 features and 50 samples. Linear Regression overfits. What should you do?",
"options": [
"Use Ridge Regression with appropriate alpha",
"Use Linear Regression without changes",
"Remove all features",
"Use logistic regression"
],
"correctAnswerIndex": 0,
"explanation": "Regularization like Ridge helps prevent overfitting when features outnumber samples."
},
{
"id": 15,
"questionText": "Scenario: Ridge Regression gives large coefficients even after standardization. Likely reason?",
"options": [
"Alpha is too small",
"Data has no noise",
"Features are uncorrelated",
"Model is perfect"
],
"correctAnswerIndex": 0,
"explanation": "A small alpha means the penalty is weak, so coefficients remain large."
},
{
"id": 16,
"questionText": "Scenario: After increasing alpha, training error increased but test error decreased. This illustrates:",
"options": [
"Bias-variance tradeoff",
"Overfitting",
"Underfitting",
"Multicollinearity"
],
"correctAnswerIndex": 0,
"explanation": "Increasing alpha increases bias (higher training error) but reduces variance (lower test error)."
},
{
"id": 17,
"questionText": "Which Python library provides Ridge Regression?",
"options": [
"numpy",
"pandas",
"scikit-learn",
"matplotlib"
],
"correctAnswerIndex": 2,
"explanation": "Scikit-learn provides Ridge regression through sklearn.linear_model.Ridge."
},
{
"id": 18,
"questionText": "Which parameter in Ridge controls regularization strength?",
"options": [
"beta",
"lambda",
"alpha",
"gamma"
],
"correctAnswerIndex": 2,
"explanation": "In scikit-learn's Ridge, alpha sets the L2 penalty strength."
},
{
"id": 19,
"questionText": "Ridge Regression reduces multicollinearity by:",
"options": [
"Shrinking correlated coefficients",
"Eliminating features",
"Adding noise",
"Creating polynomial features"
],
"correctAnswerIndex": 0,
"explanation": "L2 regularization shrinks correlated coefficients to reduce instability."
},
{
"id": 20,
"questionText": "Ridge Regression can be used for:",
"options": [
"Regression only",
"Classification only",
"Clustering",
"Principal Component Analysis"
],
"correctAnswerIndex": 0,
"explanation": "Ridge is an extension of Linear Regression and is used for regression tasks."
},
{
"id": 21,
"questionText": "Standardizing features before Ridge is important because:",
"options": [
"It reduces alpha value automatically",
"It ensures regularization treats all features equally",
"It changes the target variable",
"It creates sparse solutions"
],
"correctAnswerIndex": 1,
"explanation": "Without standardization, features with larger scales are penalized more than smaller ones."
},
{
"id": 22,
"questionText": "Scenario: Alpha is set very high. Likely effect on model?",
"options": [
"Overfitting",
"Underfitting",
"Perfect fit",
"No effect"
],
"correctAnswerIndex": 1,
"explanation": "Very high alpha over-penalizes coefficients, increasing bias and underfitting the data."
},
{
"id": 23,
"questionText": "Which type of regularization does Ridge use?",
"options": [
"L1",
"L2",
"Elastic Net",
"Dropout"
],
"correctAnswerIndex": 1,
"explanation": "Ridge uses L2 regularization to shrink coefficients."
},
{
"id": 24,
"questionText": "Scenario: Two features are highly correlated. Ridge Regression will:",
"options": [
"Randomly select one feature",
"Shrink their coefficients without eliminating either",
"Eliminate both features",
"Increase their coefficients"
],
"correctAnswerIndex": 1,
"explanation": "Ridge shrinks correlated coefficients but keeps both in the model."
},
{
"id": 25,
"questionText": "Scenario: Dataset has noisy features. Ridge Regression helps by:",
"options": [
"Ignoring noise completely",
"Reducing coefficient magnitudes to prevent overfitting",
"Removing noisy features automatically",
"Converting data to categorical"
],
"correctAnswerIndex": 1,
"explanation": "Regularization reduces sensitivity to noise, helping the model generalize better."
},
{
"id": 26,
"questionText": "Scenario: You applied Ridge Regression but your test error is still high. What could help?",
"options": [
"Decrease alpha",
"Increase alpha or try dimensionality reduction",
"Remove the intercept",
"Ignore standardization"
],
"correctAnswerIndex": 1,
"explanation": "Increasing regularization or using PCA/PLS can help improve generalization when test error is high."
},
{
"id": 27,
"questionText": "Scenario: Two datasets have the same features, but one has highly correlated inputs. Ridge Regression will:",
"options": [
"Shrink coefficients more for correlated features",
"Perform the same on both",
"Eliminate correlated features",
"Fail to converge"
],
"correctAnswerIndex": 0,
"explanation": "Ridge handles multicollinearity by shrinking coefficients of correlated features."
},
{
"id": 28,
"questionText": "How can you choose the optimal alpha in Ridge Regression?",
"options": [
"Random guess",
"Cross-validation on a range of alpha values",
"Using R-squared only",
"Using the number of features"
],
"correctAnswerIndex": 1,
"explanation": "Cross-validation is used to evaluate model performance for different alpha values and select the best one."
},
{
"id": 29,
"questionText": "Ridge Regression vs Linear Regression: which statement is true?",
"options": [
"Ridge ignores some features",
"Ridge always has lower training error",
"Ridge adds L2 penalty to reduce coefficient magnitude",
"Ridge cannot handle more features than samples"
],
"correctAnswerIndex": 2,
"explanation": "The L2 penalty in Ridge helps shrink coefficients to reduce overfitting."
},
{
"id": 30,
"questionText": "Scenario: You have standardized features and apply Ridge Regression with alpha=0.1. Increasing alpha to 10 will:",
"options": [
"Increase training error and may decrease test error",
"Decrease both training and test errors",
"Have no effect",
"Eliminate some features automatically"
],
"correctAnswerIndex": 0,
"explanation": "Higher alpha increases bias (training error) but can reduce variance (improve test error)."
},
{
"id": 31,
"questionText": "Why is Ridge Regression sensitive to feature scaling?",
"options": [
"L2 penalty depends on coefficient magnitude, which depends on feature scale",
"It uses absolute values",
"It ignores intercept",
"It only works with integers"
],
"correctAnswerIndex": 0,
"explanation": "Without scaling, large-scale features are penalized more than small-scale features."
},
{
"id": 32,
"questionText": "Scenario: You have a polynomial dataset. Ridge Regression helps by:",
"options": [
"Eliminating polynomial terms",
"Reducing overfitting caused by high-degree terms",
"Making all coefficients equal",
"Removing intercept automatically"
],
"correctAnswerIndex": 1,
"explanation": "Ridge shrinks coefficients of high-degree polynomial terms, reducing overfitting."
},
{
"id": 33,
"questionText": "Scenario: Ridge Regression and Lasso applied on same dataset. Lasso gives some zero coefficients while Ridge does not. Why?",
"options": [
"Ridge uses L1 penalty",
"Ridge uses L2 penalty which shrinks but doesn’t eliminate coefficients",
"Lasso ignores correlated features",
"Ridge ignores alpha"
],
"correctAnswerIndex": 1,
"explanation": "L2 penalty in Ridge shrinks coefficients, while L1 penalty in Lasso can set them exactly to zero."
},
{
"id": 34,
"questionText": "Scenario: Your dataset has features with very different scales. What should you do before Ridge Regression?",
"options": [
"Normalize or standardize features",
"Leave features as they are",
"Add noise to smaller features",
"Remove largest features"
],
"correctAnswerIndex": 0,
"explanation": "Standardizing ensures the penalty treats all features equally."
},
{
"id": 35,
"questionText": "Scenario: You applied Ridge Regression on noisy data. The coefficients are smaller than in Linear Regression. Why?",
"options": [
"Ridge ignores noise",
"L2 penalty shrinks coefficients to reduce overfitting",
"Noise is removed automatically",
"Training error increases"
],
"correctAnswerIndex": 1,
"explanation": "Regularization shrinks coefficients, making the model less sensitive to noise."
},
{
"id": 36,
"questionText": "Scenario: You have highly correlated features and want some coefficients exactly zero. What should you use?",
"options": [
"Ridge Regression",
"Lasso Regression",
"Linear Regression",
"Polynomial Regression"
],
"correctAnswerIndex": 1,
"explanation": "Lasso uses L1 penalty which can set some coefficients exactly to zero, performing feature selection."
},
{
"id": 37,
"questionText": "Scenario: Ridge Regression shows underfitting. What adjustment can help?",
"options": [
"Decrease alpha",
"Increase alpha",
"Remove standardization",
"Add noise"
],
"correctAnswerIndex": 0,
"explanation": "Lowering alpha reduces regularization, allowing coefficients to fit data better."
},
{
"id": 38,
"questionText": "Scenario: Two Ridge models with different alpha are trained. Model A (low alpha) has low training error, high test error. Model B (high alpha) has higher training error, lower test error. This illustrates:",
"options": [
"Bias-variance tradeoff",
"Underfitting",
"Multicollinearity",
"Polynomial expansion"
],
"correctAnswerIndex": 0,
"explanation": "Increasing alpha increases bias (higher training error) but reduces variance (better generalization)."
},
{
"id": 39,
"questionText": "Scenario: Ridge Regression on dataset with 10,000 features. Most features are irrelevant. Which is better?",
"options": [
"Ridge Regression",
"Lasso Regression",
"Standard Linear Regression",
"Decision Tree"
],
"correctAnswerIndex": 1,
"explanation": "Lasso can eliminate irrelevant features via L1 penalty, producing sparse coefficients."
},
{
"id": 40,
"questionText": "Scenario: After Ridge Regression, coefficients of correlated features are close but non-zero. This is expected because:",
"options": [
"Ridge ignores correlation",
"L2 penalty shrinks correlated coefficients equally",
"L1 penalty would do the same",
"Model is underfitting"
],
"correctAnswerIndex": 1,
"explanation": "Ridge shrinks coefficients of correlated features similarly, avoiding instability."
},
{
"id": 41,
"questionText": "Scenario: You want Ridge Regression but with some feature selection. Which method combines L1 and L2 penalties?",
"options": [
"Lasso",
"Elastic Net",
"Linear Regression",
"Polynomial Regression"
],
"correctAnswerIndex": 1,
"explanation": "Elastic Net combines L1 (feature selection) and L2 (shrinkage) penalties."
},
{
"id": 42,
"questionText": "Scenario: Ridge Regression applied without standardization. What can happen?",
"options": [
"Features with larger scale get larger penalties",
"All coefficients shrink equally",
"Training error drops",
"Alpha becomes irrelevant"
],
"correctAnswerIndex": 0,
"explanation": "Without scaling, features with larger magnitude are penalized more, biasing the model."
},
{
"id": 43,
"questionText": "Scenario: Ridge Regression applied to high-degree polynomial features. Main risk:",
"options": [
"Underfitting",
"Overfitting due to many terms",
"Alpha is too low",
"Features become sparse"
],
"correctAnswerIndex": 1,
"explanation": "High-degree polynomial features increase model complexity; Ridge shrinks coefficients to control overfitting."
},
{
"id": 44,
"questionText": "Scenario: You want to compare Ridge Regression performance with different alpha. Best approach?",
"options": [
"Single train-test split",
"K-fold cross-validation",
"Use R-squared only",
"Ignore alpha values"
],
"correctAnswerIndex": 1,
"explanation": "K-fold CV allows evaluating different alpha values reliably and selecting the optimal one."
},
{
"id": 45,
"questionText": "Scenario: Ridge Regression model has high training error and high test error. What’s happening?",
"options": [
"Underfitting due to too high alpha",
"Overfitting",
"Model perfect",
"Features irrelevant"
],
"correctAnswerIndex": 0,
"explanation": "High alpha over-penalizes coefficients, increasing bias and underfitting the data."
},
{
"id": 46,
"questionText": "Scenario: Dataset has multicollinearity. Which regression reduces variance without eliminating features?",
"options": [
"Ridge Regression",
"Lasso Regression",
"Linear Regression",
"Polynomial Regression"
],
"correctAnswerIndex": 0,
"explanation": "Ridge reduces coefficient magnitude for correlated features, lowering variance without zeroing coefficients."
},
{
"id": 47,
"questionText": "Scenario: Ridge Regression on noisy data. Coefficients are smaller than Linear Regression. Why?",
"options": [
"Noise removed automatically",
"L2 penalty shrinks coefficients",
"Model ignores target variable",
"Alpha is zero"
],
"correctAnswerIndex": 1,
"explanation": "L2 penalty makes the model less sensitive to noise by shrinking coefficients."
},
{
"id": 48,
"questionText": "Scenario: Ridge Regression applied to dataset with features on vastly different scales. Outcome?",
"options": [
"Some features penalized more than others",
"All coefficients equal",
"Alpha becomes zero",
"Model fails"
],
"correctAnswerIndex": 0,
"explanation": "Without scaling, large-scale features incur larger penalties than small-scale features."
},
{
"id": 49,
"questionText": "Scenario: Ridge Regression used for dataset with correlated inputs. What happens to their coefficients?",
"options": [
"Shrink similarly, remain non-zero",
"Zeroed out automatically",
"Become negative",
"Removed from model"
],
"correctAnswerIndex": 0,
"explanation": "Ridge shrinks correlated coefficients together without eliminating them."
},
{
"id": 50,
"questionText": "Scenario: You need Ridge Regression but also want feature selection. Best choice?",
"options": [
"Increase alpha",
"Use Elastic Net combining L1 and L2",
"Decrease alpha",
"Ignore multicollinearity"
],
"correctAnswerIndex": 1,
"explanation": "Elastic Net allows both shrinkage (L2) and feature selection (L1)."
},
{
"id": 51,
"questionText": "Scenario: Ridge Regression is applied to a dataset with 5000 features, most of which are correlated. What is the main advantage?",
"options": [
"Eliminates irrelevant features",
"Reduces coefficient variance without removing features",
"Always decreases bias to zero",
"Removes noise automatically"
],
"correctAnswerIndex": 1,
"explanation": "Ridge shrinks correlated feature coefficients to reduce variance, maintaining all features in the model."
},
{
"id": 52,
"questionText": "Scenario: After Ridge Regression, test error is still high. Possible solution?",
"options": [
"Increase alpha further",
"Use dimensionality reduction like PCA before Ridge",
"Remove standardization",
"Reduce training samples"
],
"correctAnswerIndex": 1,
"explanation": "Dimensionality reduction can remove redundant features and improve generalization."
},
{
"id": 53,
"questionText": "Scenario: Ridge Regression applied to dataset with polynomial features. Observed very high coefficients for high-degree terms. Best approach?",
"options": [
"Increase alpha",
"Decrease alpha",
"Remove intercept",
"Ignore polynomial terms"
],
"correctAnswerIndex": 0,
"explanation": "Increasing alpha penalizes large coefficients, controlling overfitting in polynomial terms."
},
{
"id": 54,
"questionText": "Scenario: Ridge Regression on dataset with noisy inputs and high multicollinearity. Observed stable coefficients. Why?",
"options": [
"L2 penalty reduces sensitivity to noise and stabilizes correlated coefficients",
"Training error is minimized",
"Alpha is zero",
"Model ignores correlated features"
],
"correctAnswerIndex": 0,
"explanation": "Ridge shrinks coefficients to stabilize model against noise and multicollinearity."
},
{
"id": 55,
"questionText": "Scenario: You perform Ridge Regression with alpha=1 and 10-fold cross-validation. Best alpha is found to be 5. Interpretation?",
"options": [
"Model underfits with alpha=1, alpha=5 improves generalization",
"Model overfits with alpha=5",
"Cross-validation is irrelevant",
"Training error is minimal at alpha=1"
],
"correctAnswerIndex": 0,
"explanation": "Higher alpha increases bias slightly but reduces variance, improving test performance."
},
{
"id": 56,
"questionText": "Scenario: Ridge Regression applied to standardized features. Coefficients of two correlated features are nearly equal. This occurs because:",
"options": [
"Alpha is too high",
"L2 penalty shrinks correlated coefficients similarly",
"Features are independent",
"Standardization is not needed"
],
"correctAnswerIndex": 1,
"explanation": "Ridge shrinks correlated coefficients together, leading to similar values."
},
{
"id": 57,
"questionText": "Scenario: You applied Ridge Regression with alpha=0. Ridge behaves like:",
"options": [
"Lasso Regression",
"Linear Regression",
"Elastic Net",
"Polynomial Regression"
],
"correctAnswerIndex": 1,
"explanation": "Alpha=0 removes the L2 penalty, reducing Ridge to standard Linear Regression."
},
{
"id": 58,
"questionText": "Scenario: Ridge Regression applied on dataset with 1000 features, many irrelevant. Which method could improve sparsity?",
"options": [
"Increase alpha",
"Switch to Lasso or Elastic Net",
"Decrease alpha",
"Use standardization only"
],
"correctAnswerIndex": 1,
"explanation": "Lasso (L1) or Elastic Net can set irrelevant coefficients to zero, creating sparse models."
},
{
"id": 59,
"questionText": "Scenario: Ridge Regression used for a dataset with missing values. Best approach?",
"options": [
"Ridge handles missing automatically",
"Impute missing values before applying Ridge",
"Remove alpha",
"Ignore missing values"
],
"correctAnswerIndex": 1,
"explanation": "Ridge requires complete data; missing values should be imputed or removed first."
},
{
"id": 60,
"questionText": "Scenario: Ridge Regression on standardized dataset shows training error slightly higher than Linear Regression but test error lower. Reason?",
"options": [
"Bias-variance tradeoff: Ridge increased bias slightly but reduced variance",
"Model underfits completely",
"Alpha is zero",
"Data is too small"
],
"correctAnswerIndex": 0,
"explanation": "Regularization increases bias but reduces variance, improving test performance."
},
{
"id": 61,
"questionText": "Scenario: Ridge Regression applied with high alpha and low alpha. Observed training and test errors: High alpha → high training, low test; Low alpha → low training, high test. This illustrates:",
"options": [
"Bias-variance tradeoff",
"Overfitting",
"Multicollinearity",
"Polynomial expansion"
],
"correctAnswerIndex": 0,
"explanation": "This is the classic bias-variance tradeoff scenario."
},
{
"id": 62,
"questionText": "Scenario: Ridge Regression on dataset with categorical variables encoded as one-hot vectors. Main concern?",
"options": [
"High multicollinearity due to dummy variables",
"Alpha selection irrelevant",
"Scaling not required",
"Target variable changes"
],
"correctAnswerIndex": 0,
"explanation": "One-hot encoding can produce correlated dummy features; Ridge helps reduce coefficient variance."
},
{
"id": 63,
"questionText": "Scenario: Ridge Regression applied on a time-series dataset with lag features. Why standardization is important?",
"options": [
"Alpha only applies to standardized features",
"Regularization penalizes coefficients fairly only if features are on same scale",
"Intercept is ignored otherwise",
"Time index must be normalized"
],
"correctAnswerIndex": 1,
"explanation": "L2 penalty shrinks coefficients fairly only when all features are standardized."
},
{
"id": 64,
"questionText": "Scenario: Ridge Regression and OLS applied on small dataset with multicollinearity. Observed unstable coefficients with OLS, stable with Ridge. Why?",
"options": [
"Ridge reduces coefficient variance through regularization",
"Ridge increases training error",
"OLS ignores multicollinearity",
"Alpha is zero"
],
"correctAnswerIndex": 0,
"explanation": "Regularization stabilizes coefficients for correlated features."
},
{
"id": 65,
"questionText": "Scenario: Ridge Regression applied after PCA. Advantage?",
"options": [
"Reduces dimensionality, coefficients shrunk on principal components",
"Eliminates intercept",
"No need for alpha",
"Features become sparse"
],
"correctAnswerIndex": 0,
"explanation": "PCA reduces dimensionality; Ridge shrinks coefficients on principal components to control overfitting."
},
{
"id": 66,
"questionText": "Scenario: Ridge Regression applied with alpha very large. Observed training and test errors both high. Reason?",
"options": [
"Underfitting due to over-penalization",
"Overfitting",
"Alpha too small",
"Data not standardized"
],
"correctAnswerIndex": 0,
"explanation": "Excessive alpha causes high bias, leading to underfitting."
},
{
"id": 67,
"questionText": "Scenario: Ridge Regression applied to polynomial regression with degree 10. Why use Ridge?",
"options": [
"Prevent overfitting from high-degree polynomial terms",
"Increase training error",
"Eliminate low-degree terms",
"Remove intercept automatically"
],
"correctAnswerIndex": 0,
"explanation": "Ridge penalizes large coefficients from high-degree terms, reducing overfitting."
},
{
"id": 68,
"questionText": "Scenario: Ridge Regression applied to features with different units. Observed coefficients of small-scale features larger than large-scale ones. Reason?",
"options": [
"L2 penalty uneven due to lack of standardization",
"Alpha is zero",
"Model overfits",
"Data too small"
],
"correctAnswerIndex": 0,
"explanation": "Without standardization, penalty is unfair; features with small scale are penalized less."
},
{
"id": 69,
"questionText": "Scenario: Ridge Regression applied with cross-validation. Optimal alpha selected minimizes:",
"options": [
"Training error only",
"Test error on validation folds",
"Number of features",
"Intercept value"
],
"correctAnswerIndex": 1,
"explanation": "Cross-validation selects alpha that minimizes validation/test error, improving generalization."
},
{
"id": 70,
"questionText": "Scenario: Ridge Regression applied to dataset with multicollinearity. Coefficients are shrunk but all non-zero. Implication?",
"options": [
"Variance reduced, all features retained",
"Model overfits",
"Features eliminated automatically",
"Model fails"
],
"correctAnswerIndex": 0,
"explanation": "Ridge reduces variance without eliminating correlated features."
},
{
"id": 71,
"questionText": "Scenario: Ridge Regression applied to large-scale dataset. Alpha tuning via grid search. Why important?",
"options": [
"Different alphas balance bias and variance for optimal performance",
"Alpha irrelevant",
"Alpha only affects training error",
"Regularization unnecessary"
],
"correctAnswerIndex": 0,
"explanation": "Alpha controls regularization strength; tuning balances bias and variance."
},
{
"id": 72,
"questionText": "Scenario: Ridge Regression applied with very small alpha. Observed high variance. Why?",
"options": [
"L2 penalty too weak to control overfitting",
"Alpha too large",
"Data unstandardized",
"Intercept ignored"
],
"correctAnswerIndex": 0,
"explanation": "Small alpha provides minimal regularization, leaving high-variance coefficients unchecked."
},
{
"id": 73,
"questionText": "Scenario: Ridge Regression vs Lasso on highly correlated features. Expected result?",
"options": [
"Ridge shrinks coefficients similarly; Lasso selects one and zeroes others",
"Both eliminate all correlated features",
"Ridge produces sparse solution; Lasso does not",
"Alpha irrelevant"
],
"correctAnswerIndex": 0,
"explanation": "Ridge keeps all correlated features with smaller coefficients; Lasso may zero some."
},
{
"id": 74,
"questionText": "Scenario: Ridge Regression applied with alpha=0. Model behaves like:",
"options": [
"Linear Regression",
"Lasso",
"Elastic Net",
"Polynomial Regression"
],
"correctAnswerIndex": 0,
"explanation": "Alpha=0 removes L2 penalty, reducing Ridge to standard Linear Regression."
},
{
"id": 75,
"questionText": "Scenario: Ridge Regression applied to a dataset with 1000 features, some irrelevant. How to reduce irrelevant features?",
"options": [
"Switch to Lasso or Elastic Net",
"Increase alpha excessively",
"Decrease alpha to zero",
"Ignore standardization"
],
"correctAnswerIndex": 0,
"explanation": "Lasso or Elastic Net can remove irrelevant features via L1 regularization."
},
{
"id": 76,
"questionText": "Scenario: Ridge Regression applied on medical dataset with 500 features, many correlated. Goal: predict patient outcome. Best approach?",
"options": [
"Use Ridge Regression with cross-validated alpha",
"Use standard Linear Regression",
"Use Lasso only",
"Ignore multicollinearity"
],
"correctAnswerIndex": 0,
"explanation": "Ridge handles correlated features effectively and cross-validation selects optimal alpha."
},
{
"id": 77,
"questionText": "Scenario: Ridge Regression applied to dataset with outliers. Observation: coefficients not extremely affected. Why?",
"options": [
"L2 penalty shrinks coefficients, reducing sensitivity to outliers",
"Training error minimized",
"Model ignores target variable",
"Alpha is zero"
],
"correctAnswerIndex": 0,
"explanation": "Regularization prevents large coefficients, making the model less sensitive to outliers."
},
{
"id": 78,
"questionText": "Scenario: Ridge Regression applied with alpha very small, results similar to Linear Regression. Interpretation?",
"options": [
"L2 penalty is too weak to control overfitting",
"Model underfits",
"Coefficients are zero",
"Training error very high"
],
"correctAnswerIndex": 0,
"explanation": "Small alpha means minimal regularization; Ridge behaves like Linear Regression with potential overfitting."
},
{
"id": 79,
"questionText": "Scenario: Ridge Regression on dataset with 2000 features, many irrelevant. Test error high. Recommended?",
"options": [
"Switch to Lasso or Elastic Net",
"Increase alpha excessively",
"Decrease alpha",
"Ignore irrelevant features"
],
"correctAnswerIndex": 0,
"explanation": "Lasso or Elastic Net can remove irrelevant features, improving generalization."
},
{
"id": 80,
"questionText": "Scenario: Ridge Regression applied on highly noisy dataset. Observed smaller coefficients than Linear Regression. Why?",
"options": [
"L2 penalty shrinks coefficients to reduce variance",
"Model ignores noise",
"Training error drops to zero",
"Alpha is zero"
],
"correctAnswerIndex": 0,
"explanation": "Regularization reduces sensitivity to noise, shrinking coefficients."
},
{
"id": 81,
"questionText": "Scenario: Ridge Regression applied to polynomial features with high degree. Observation: large coefficients for high-degree terms. Best solution?",
"options": [
"Increase alpha to penalize large coefficients",
"Decrease alpha",
"Remove intercept",
"Ignore polynomial terms"
],
"correctAnswerIndex": 0,
"explanation": "Higher alpha controls overfitting from high-degree polynomial terms."
},
{
"id": 82,
"questionText": "Scenario: Ridge Regression applied to dataset with categorical features encoded as one-hot vectors. Concern?",
"options": [
"Multicollinearity due to dummy variables",
"Alpha irrelevant",
"Scaling not required",
"Target variable changes"
],
"correctAnswerIndex": 0,
"explanation": "One-hot encoding creates correlated dummy features; Ridge shrinks their coefficients."
},
{
"id": 83,
"questionText": "Scenario: Ridge Regression applied to time-series dataset with lag features. Why standardization important?",
"options": [
"L2 penalty penalizes coefficients fairly only if features are on same scale",
"Intercept ignored otherwise",
"Time index must be normalized",
"Alpha irrelevant"
],
"correctAnswerIndex": 0,
"explanation": "Standardizing features ensures L2 penalty treats all lag features fairly."
},
{
"id": 84,
"questionText": "Scenario: Ridge Regression applied on dataset with missing values. Action required?",
"options": [
"Impute missing values before applying Ridge",
"Remove alpha",
"Ignore missing values",
"L2 penalty handles missing automatically"
],
"correctAnswerIndex": 0,
"explanation": "Ridge requires complete data; missing values must be imputed first."
},
{
"id": 85,
"questionText": "Scenario: Ridge Regression with very high alpha. Observed high training and test errors. Reason?",
"options": [
"Underfitting due to over-penalization",
"Overfitting",
"Alpha too small",
"Data not standardized"
],
"correctAnswerIndex": 0,
"explanation": "Excessive alpha increases bias, causing underfitting."
},
{
"id": 86,
"questionText": "Scenario: Ridge Regression applied on dataset with highly correlated features. Coefficients shrunk but non-zero. Implication?",
"options": [
"Variance reduced, features retained",
"Overfitting",
"Features eliminated",
"Model fails"
],
"correctAnswerIndex": 0,
"explanation": "Ridge reduces variance without removing correlated features."
},
{
"id": 87,
"questionText": "Scenario: Ridge Regression applied on large-scale dataset. Why tune alpha via grid search?",
"options": [
"Alpha balances bias-variance tradeoff for optimal performance",
"Alpha irrelevant",
"Alpha only affects training error",
"Regularization unnecessary"
],
"correctAnswerIndex": 0,
"explanation": "Grid search finds alpha that provides the best tradeoff between bias and variance."
},
{
"id": 88,
"questionText": "Scenario: Ridge Regression applied to dataset with polynomial features. High-degree terms dominate coefficients. Solution?",
"options": [
"Increase alpha to control large coefficients",
"Decrease alpha",
"Remove polynomial terms",
"Ignore coefficients"
],
"correctAnswerIndex": 0,
"explanation": "Higher alpha penalizes large coefficients, reducing overfitting."
},
{
"id": 89,
"questionText": "Scenario: Ridge Regression applied on dataset with features in different units. Observation: large coefficients for small-scale features. Reason?",
"options": [
"L2 penalty uneven due to lack of standardization",
"Alpha too high",
"Data uncorrelated",
"Training error minimal"
],
"correctAnswerIndex": 0,
"explanation": "Without standardization, small-scale features are penalized less, leading to larger coefficients."
},
{
"id": 90,
"questionText": "Scenario: Ridge Regression applied with k-fold cross-validation. Optimal alpha minimizes:",
"options": [
"Validation/test error",
"Training error",
"Number of features",
"Intercept value"
],
"correctAnswerIndex": 0,
"explanation": "Cross-validation selects alpha that minimizes test error for better generalization."
},
{
"id": 91,
"questionText": "Scenario: Ridge Regression applied with very small alpha. Observed high variance. Reason?",
"options": [
"L2 penalty too weak to control overfitting",
"Alpha too large",
"Data unstandardized",
"Intercept ignored"
],
"correctAnswerIndex": 0,
"explanation": "Minimal regularization leaves coefficients unchecked, causing high variance."
},
{
"id": 92,
"questionText": "Scenario: Ridge Regression applied alongside Lasso on same dataset. Expected difference?",
"options": [
"Ridge shrinks coefficients; Lasso may zero some",
"Both produce sparse solutions",
"Ridge eliminates features; Lasso does not",
"Alpha irrelevant"
],
"correctAnswerIndex": 0,
"explanation": "Ridge keeps all coefficients small but non-zero; Lasso can perform feature selection."
},
{
"id": 93,
"questionText": "Scenario: Ridge Regression applied to dataset with irrelevant features. Test error high. Solution?",
"options": [
"Switch to Lasso or Elastic Net",
"Increase alpha excessively",
"Decrease alpha",
"Ignore irrelevant features"
],
"correctAnswerIndex": 0,
"explanation": "Lasso or Elastic Net can remove irrelevant features to improve performance."
},
{
"id": 94,
"questionText": "Scenario: Ridge Regression applied with standardized features. Coefficients for correlated features similar. Reason?",
"options": [
"L2 penalty shrinks correlated coefficients similarly",
"Alpha too low",
"Features independent",
"Data too small"
],
"correctAnswerIndex": 0,
"explanation": "Ridge shrinks correlated coefficients together, producing similar values."
},
{
"id": 95,
"questionText": "Scenario: Ridge Regression applied to dataset with noisy features. Coefficients smaller than Linear Regression. Why?",
"options": [
"Regularization reduces sensitivity to noise",
"Training error minimized",
"Alpha zero",
"Noise ignored"
],
"correctAnswerIndex": 0,
"explanation": "L2 penalty shrinks coefficients, making model less sensitive to noise."
},
{
"id": 96,
"questionText": "Scenario: Ridge Regression applied to polynomial regression of degree 12. High-degree terms produce large coefficients. Solution?",
"options": [
"Increase alpha",
"Decrease alpha",
"Remove intercept",
"Ignore coefficients"
],
"correctAnswerIndex": 0,
"explanation": "Higher alpha controls overfitting by shrinking large coefficients from high-degree terms."
},
{
"id": 97,
"questionText": "Scenario: Ridge Regression applied on dataset with one-hot encoded features. Concern?",
"options": [
"Multicollinearity due to dummy variables",
"Alpha irrelevant",
"Scaling not needed",
"Intercept ignored"
],
"correctAnswerIndex": 0,
"explanation": "One-hot encoding creates correlated dummy variables; Ridge shrinks their coefficients."
},
{
"id": 98,
"questionText": "Scenario: Ridge Regression applied on dataset with missing values. Action required?",
"options": [
"Impute missing values first",
"Ignore missing values",
"Remove alpha",
"L2 penalty handles missing automatically"
],
"correctAnswerIndex": 0,
"explanation": "Ridge requires complete data; missing values must be imputed or removed."
},
{
"id": 99,
"questionText": "Scenario: Ridge Regression applied with cross-validation. Selected alpha minimizes:",
"options": [
"Validation/test error",
"Training error only",
"Number of features",
"Intercept value"
],
"correctAnswerIndex": 0,
"explanation": "Cross-validation selects alpha that minimizes validation error for optimal generalization."
},
{
"id": 100,
"questionText": "Scenario: Ridge Regression applied to real-world dataset with high multicollinearity, noisy features, and high-dimensionality. Best approach?",
"options": [
"Standardize features, tune alpha via cross-validation, consider Elastic Net if feature selection needed",
"Use Linear Regression",
"Ignore alpha",
"Remove L2 penalty"
],
"correctAnswerIndex": 0,
"explanation": "Standardization and cross-validated Ridge handle noise and multicollinearity; Elastic Net adds feature selection."
}
]
}