| { |
| "title": "XGBoost Regression Mastery: 100 MCQs", |
| "description": "A complete set of 100 multiple-choice questions covering XGBoost Regression \u2014 from basics to advanced tuning and scenario-based problem solving.", |
| "questions": [ |
| { |
| "id": 1, |
| "questionText": "What does XGBoost primarily stand for in machine learning?", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 3, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 2, |
| "questionText": "What does XGBoost primarily stand for in machine learning?", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 0, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 3, |
| "questionText": "What does XGBoost primarily stand for in machine learning?", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 0, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 4, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q4)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 5, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q5)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 3, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 6, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q6)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 7, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q7)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 8, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q8)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 9, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q9)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 0, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 10, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q10)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 1, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 11, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q11)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 0, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 12, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q12)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 3, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 13, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q13)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 3, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 14, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q14)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 1, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 15, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q15)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 1, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 16, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q16)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 3, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 17, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q17)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 18, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q18)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 19, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q19)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 3, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 20, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q20)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 21, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q21)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 0, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 22, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q22)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 0, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 23, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q23)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 3, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 24, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q24)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 3, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 25, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q25)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 1, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 26, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q26)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 27, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q27)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 28, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q28)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 1, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 29, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q29)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 30, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q30)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 0, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 31, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q31)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 1, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 32, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q32)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 0, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 33, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q33)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 34, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q34)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 35, |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q35)", |
| "options": [ |
| "Extreme Gradient Boosting", |
| "Extra Gaussian Boost", |
| "Extended Gradient Binary Output", |
| "Exponential Gain Booster" |
| ], |
| "correctAnswerIndex": 0, |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." |
| }, |
| { |
| "id": 36, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q36)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 1, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 37, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q37)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 0, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 38, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q38)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 3, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 39, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q39)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 0, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 40, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q40)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 41, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q41)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 42, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q42)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 3, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 43, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q43)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 3, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 44, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q44)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 1, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 45, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q45)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 0, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 46, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q46)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 0, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 47, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q47)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 1, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 48, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q48)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 49, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q49)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 0, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 50, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q50)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 3, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 51, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q51)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 0, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 52, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q52)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 3, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 53, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q53)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 3, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 54, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q54)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 3, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 55, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q55)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 0, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 56, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q56)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 57, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q57)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 1, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 58, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q58)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 1, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 59, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q59)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 1, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 60, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q60)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 0, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 61, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q61)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 62, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q62)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 3, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 63, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q63)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 64, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q64)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 0, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 65, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q65)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 66, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q66)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 0, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 67, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q67)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 3, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 68, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q68)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 1, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 69, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q69)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 3, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 70, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q70)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 71, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q71)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 72, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q72)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 73, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q73)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 1, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 74, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q74)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 3, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 75, |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q75)", |
| "options": [ |
| "It controls the depth of trees.", |
| "It prevents overfitting by reducing feature importance.", |
| "It scales the contribution of each tree during training.", |
| "It increases the randomness in sampling data." |
| ], |
| "correctAnswerIndex": 3, |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." |
| }, |
| { |
| "id": 76, |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q76)", |
| "options": [ |
| "Increase subsample and colsample_bytree values", |
| "Reduce regularization parameters lambda and alpha", |
| "Add more trees to the ensemble", |
| "Set booster to 'gblinear'" |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." |
| }, |
| { |
| "id": 77, |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q77)", |
| "options": [ |
| "Increase subsample and colsample_bytree values", |
| "Reduce regularization parameters lambda and alpha", |
| "Add more trees to the ensemble", |
| "Set booster to 'gblinear'" |
| ], |
| "correctAnswerIndex": 1, |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." |
| }, |
| { |
| "id": 78, |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q78)", |
| "options": [ |
| "Increase subsample and colsample_bytree values", |
| "Reduce regularization parameters lambda and alpha", |
| "Add more trees to the ensemble", |
| "Set booster to 'gblinear'" |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." |
| }, |
| { |
| "id": 79, |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q79)", |
| "options": [ |
| "Increase subsample and colsample_bytree values", |
| "Reduce regularization parameters lambda and alpha", |
| "Add more trees to the ensemble", |
| "Set booster to 'gblinear'" |
| ], |
| "correctAnswerIndex": 3, |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." |
| }, |
| { |
| "id": 80, |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q80)", |
| "options": [ |
| "Increase subsample and colsample_bytree values", |
| "Reduce regularization parameters lambda and alpha", |
| "Add more trees to the ensemble", |
| "Set booster to 'gblinear'" |
| ], |
| "correctAnswerIndex": 1, |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." |
| }, |
| { |
| "id": 81, |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q81)", |
| "options": [ |
| "Increase subsample and colsample_bytree values", |
| "Reduce regularization parameters lambda and alpha", |
| "Add more trees to the ensemble", |
| "Set booster to 'gblinear'" |
| ], |
| "correctAnswerIndex": 3, |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." |
| }, |
| { |
| "id": 82, |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q82)", |
| "options": [ |
| "Increase subsample and colsample_bytree values", |
| "Reduce regularization parameters lambda and alpha", |
| "Add more trees to the ensemble", |
| "Set booster to 'gblinear'" |
| ], |
| "correctAnswerIndex": 1, |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." |
| }, |
| { |
| "id": 83, |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q83)", |
| "options": [ |
| "Increase subsample and colsample_bytree values", |
| "Reduce regularization parameters lambda and alpha", |
| "Add more trees to the ensemble", |
| "Set booster to 'gblinear'" |
| ], |
| "correctAnswerIndex": 1, |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." |
| }, |
| { |
| "id": 84, |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q84)", |
| "options": [ |
| "Increase subsample and colsample_bytree values", |
| "Reduce regularization parameters lambda and alpha", |
| "Add more trees to the ensemble", |
| "Set booster to 'gblinear'" |
| ], |
| "correctAnswerIndex": 1, |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." |
| }, |
| { |
| "id": 85, |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q85)", |
| "options": [ |
| "Increase subsample and colsample_bytree values", |
| "Reduce regularization parameters lambda and alpha", |
| "Add more trees to the ensemble", |
| "Set booster to 'gblinear'" |
| ], |
| "correctAnswerIndex": 1, |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." |
| }, |
| { |
| "id": 86, |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q86)", |
| "options": [ |
| "Increase subsample and colsample_bytree values", |
| "Reduce regularization parameters lambda and alpha", |
| "Add more trees to the ensemble", |
| "Set booster to 'gblinear'" |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." |
| }, |
| { |
| "id": 87, |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q87)", |
| "options": [ |
| "Increase subsample and colsample_bytree values", |
| "Reduce regularization parameters lambda and alpha", |
| "Add more trees to the ensemble", |
| "Set booster to 'gblinear'" |
| ], |
| "correctAnswerIndex": 0, |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." |
| }, |
| { |
| "id": 88, |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q88)", |
| "options": [ |
| "Increase subsample and colsample_bytree values", |
| "Reduce regularization parameters lambda and alpha", |
| "Add more trees to the ensemble", |
| "Set booster to 'gblinear'" |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." |
| }, |
| { |
| "id": 89, |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q89)", |
| "options": [ |
| "Increase subsample and colsample_bytree values", |
| "Reduce regularization parameters lambda and alpha", |
| "Add more trees to the ensemble", |
| "Set booster to 'gblinear'" |
| ], |
| "correctAnswerIndex": 1, |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." |
| }, |
| { |
| "id": 90, |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q90)", |
| "options": [ |
| "Increase subsample and colsample_bytree values", |
| "Reduce regularization parameters lambda and alpha", |
| "Add more trees to the ensemble", |
| "Set booster to 'gblinear'" |
| ], |
| "correctAnswerIndex": 0, |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." |
| }, |
| { |
| "id": 91, |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q91)", |
| "options": [ |
| "Increase subsample and colsample_bytree values", |
| "Reduce regularization parameters lambda and alpha", |
| "Add more trees to the ensemble", |
| "Set booster to 'gblinear'" |
| ], |
| "correctAnswerIndex": 1, |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." |
| }, |
| { |
| "id": 92, |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q92)", |
| "options": [ |
| "Increase subsample and colsample_bytree values", |
| "Reduce regularization parameters lambda and alpha", |
| "Add more trees to the ensemble", |
| "Set booster to 'gblinear'" |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." |
| }, |
| { |
| "id": 93, |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q93)", |
| "options": [ |
| "Increase subsample and colsample_bytree values", |
| "Reduce regularization parameters lambda and alpha", |
| "Add more trees to the ensemble", |
| "Set booster to 'gblinear'" |
| ], |
| "correctAnswerIndex": 0, |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." |
| }, |
| { |
| "id": 94, |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q94)", |
| "options": [ |
| "Increase subsample and colsample_bytree values", |
| "Reduce regularization parameters lambda and alpha", |
| "Add more trees to the ensemble", |
| "Set booster to 'gblinear'" |
| ], |
| "correctAnswerIndex": 1, |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." |
| }, |
| { |
| "id": 95, |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q95)", |
| "options": [ |
| "Increase subsample and colsample_bytree values", |
| "Reduce regularization parameters lambda and alpha", |
| "Add more trees to the ensemble", |
| "Set booster to 'gblinear'" |
| ], |
| "correctAnswerIndex": 1, |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." |
| }, |
| { |
| "id": 96, |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q96)", |
| "options": [ |
| "Increase subsample and colsample_bytree values", |
| "Reduce regularization parameters lambda and alpha", |
| "Add more trees to the ensemble", |
| "Set booster to 'gblinear'" |
| ], |
| "correctAnswerIndex": 0, |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." |
| }, |
| { |
| "id": 97, |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q97)", |
| "options": [ |
| "Increase subsample and colsample_bytree values", |
| "Reduce regularization parameters lambda and alpha", |
| "Add more trees to the ensemble", |
| "Set booster to 'gblinear'" |
| ], |
| "correctAnswerIndex": 3, |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." |
| }, |
| { |
| "id": 98, |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q98)", |
| "options": [ |
| "Increase subsample and colsample_bytree values", |
| "Reduce regularization parameters lambda and alpha", |
| "Add more trees to the ensemble", |
| "Set booster to 'gblinear'" |
| ], |
| "correctAnswerIndex": 1, |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." |
| }, |
| { |
| "id": 99, |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q99)", |
| "options": [ |
| "Increase subsample and colsample_bytree values", |
| "Reduce regularization parameters lambda and alpha", |
| "Add more trees to the ensemble", |
| "Set booster to 'gblinear'" |
| ], |
| "correctAnswerIndex": 2, |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." |
| }, |
| { |
| "id": 100, |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q100)", |
| "options": [ |
| "Increase subsample and colsample_bytree values", |
| "Reduce regularization parameters lambda and alpha", |
| "Add more trees to the ensemble", |
| "Set booster to 'gblinear'" |
| ], |
| "correctAnswerIndex": 1, |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." |
| } |
| ] |
| } |