Spaces:
Build error
Build error
| { | |
| "title": "XGBoost Regression Mastery: 100 MCQs", | |
| "description": "A complete set of 100 multiple-choice questions covering XGBoost Regression \u2014 from basics to advanced tuning and scenario-based problem solving.", | |
| "questions": [ | |
| { | |
| "id": 1, | |
| "questionText": "What does XGBoost primarily stand for in machine learning?", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 3, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 2, | |
| "questionText": "What does XGBoost primarily stand for in machine learning?", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 0, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 3, | |
| "questionText": "What does XGBoost primarily stand for in machine learning?", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 0, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 4, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q4)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 5, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q5)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 3, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 6, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q6)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 7, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q7)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 8, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q8)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 9, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q9)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 0, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 10, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q10)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 1, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 11, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q11)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 0, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 12, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q12)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 3, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 13, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q13)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 3, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 14, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q14)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 1, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 15, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q15)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 1, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 16, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q16)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 3, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 17, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q17)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 18, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q18)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 19, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q19)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 3, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 20, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q20)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 21, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q21)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 0, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 22, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q22)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 0, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 23, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q23)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 3, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 24, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q24)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 3, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 25, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q25)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 1, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 26, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q26)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 27, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q27)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 28, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q28)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 1, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 29, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q29)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 30, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q30)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 0, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 31, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q31)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 1, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 32, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q32)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 0, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 33, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q33)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 34, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q34)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 35, | |
| "questionText": "What does XGBoost primarily stand for in machine learning (Easy Q35)", | |
| "options": [ | |
| "Extreme Gradient Boosting", | |
| "Extra Gaussian Boost", | |
| "Extended Gradient Binary Output", | |
| "Exponential Gain Booster" | |
| ], | |
| "correctAnswerIndex": 0, | |
| "explanation": "XGBoost stands for Extreme Gradient Boosting, a highly efficient and scalable implementation of gradient boosted decision trees." | |
| }, | |
| { | |
| "id": 36, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q36)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 1, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 37, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q37)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 0, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 38, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q38)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 3, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 39, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q39)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 0, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 40, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q40)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 41, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q41)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 42, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q42)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 3, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 43, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q43)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 3, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 44, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q44)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 1, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 45, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q45)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 0, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 46, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q46)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 0, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 47, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q47)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 1, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 48, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q48)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 49, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q49)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 0, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 50, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q50)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 3, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 51, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q51)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 0, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 52, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q52)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 3, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 53, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q53)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 3, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 54, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q54)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 3, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 55, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q55)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 0, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 56, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q56)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 57, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q57)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 1, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 58, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q58)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 1, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 59, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q59)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 1, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 60, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q60)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 0, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 61, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q61)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 62, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q62)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 3, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 63, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q63)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 64, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q64)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 0, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 65, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q65)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 66, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q66)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 0, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 67, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q67)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 3, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 68, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q68)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 1, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 69, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q69)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 3, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 70, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q70)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 71, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q71)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 72, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q72)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 73, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q73)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 1, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 74, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q74)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 3, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 75, | |
| "questionText": "In XGBoost Regression, what is the purpose of the learning rate (eta) (Medium Q75)", | |
| "options": [ | |
| "It controls the depth of trees.", | |
| "It prevents overfitting by reducing feature importance.", | |
| "It scales the contribution of each tree during training.", | |
| "It increases the randomness in sampling data." | |
| ], | |
| "correctAnswerIndex": 3, | |
| "explanation": "The learning rate (eta) determines how much each new tree contributes to the overall model. Smaller values make learning slower but often improve generalization." | |
| }, | |
| { | |
| "id": 76, | |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q76)", | |
| "options": [ | |
| "Increase subsample and colsample_bytree values", | |
| "Reduce regularization parameters lambda and alpha", | |
| "Add more trees to the ensemble", | |
| "Set booster to 'gblinear'" | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." | |
| }, | |
| { | |
| "id": 77, | |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q77)", | |
| "options": [ | |
| "Increase subsample and colsample_bytree values", | |
| "Reduce regularization parameters lambda and alpha", | |
| "Add more trees to the ensemble", | |
| "Set booster to 'gblinear'" | |
| ], | |
| "correctAnswerIndex": 1, | |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." | |
| }, | |
| { | |
| "id": 78, | |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q78)", | |
| "options": [ | |
| "Increase subsample and colsample_bytree values", | |
| "Reduce regularization parameters lambda and alpha", | |
| "Add more trees to the ensemble", | |
| "Set booster to 'gblinear'" | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." | |
| }, | |
| { | |
| "id": 79, | |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q79)", | |
| "options": [ | |
| "Increase subsample and colsample_bytree values", | |
| "Reduce regularization parameters lambda and alpha", | |
| "Add more trees to the ensemble", | |
| "Set booster to 'gblinear'" | |
| ], | |
| "correctAnswerIndex": 3, | |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." | |
| }, | |
| { | |
| "id": 80, | |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q80)", | |
| "options": [ | |
| "Increase subsample and colsample_bytree values", | |
| "Reduce regularization parameters lambda and alpha", | |
| "Add more trees to the ensemble", | |
| "Set booster to 'gblinear'" | |
| ], | |
| "correctAnswerIndex": 1, | |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." | |
| }, | |
| { | |
| "id": 81, | |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q81)", | |
| "options": [ | |
| "Increase subsample and colsample_bytree values", | |
| "Reduce regularization parameters lambda and alpha", | |
| "Add more trees to the ensemble", | |
| "Set booster to 'gblinear'" | |
| ], | |
| "correctAnswerIndex": 3, | |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." | |
| }, | |
| { | |
| "id": 82, | |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q82)", | |
| "options": [ | |
| "Increase subsample and colsample_bytree values", | |
| "Reduce regularization parameters lambda and alpha", | |
| "Add more trees to the ensemble", | |
| "Set booster to 'gblinear'" | |
| ], | |
| "correctAnswerIndex": 1, | |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." | |
| }, | |
| { | |
| "id": 83, | |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q83)", | |
| "options": [ | |
| "Increase subsample and colsample_bytree values", | |
| "Reduce regularization parameters lambda and alpha", | |
| "Add more trees to the ensemble", | |
| "Set booster to 'gblinear'" | |
| ], | |
| "correctAnswerIndex": 1, | |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." | |
| }, | |
| { | |
| "id": 84, | |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q84)", | |
| "options": [ | |
| "Increase subsample and colsample_bytree values", | |
| "Reduce regularization parameters lambda and alpha", | |
| "Add more trees to the ensemble", | |
| "Set booster to 'gblinear'" | |
| ], | |
| "correctAnswerIndex": 1, | |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." | |
| }, | |
| { | |
| "id": 85, | |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q85)", | |
| "options": [ | |
| "Increase subsample and colsample_bytree values", | |
| "Reduce regularization parameters lambda and alpha", | |
| "Add more trees to the ensemble", | |
| "Set booster to 'gblinear'" | |
| ], | |
| "correctAnswerIndex": 1, | |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." | |
| }, | |
| { | |
| "id": 86, | |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q86)", | |
| "options": [ | |
| "Increase subsample and colsample_bytree values", | |
| "Reduce regularization parameters lambda and alpha", | |
| "Add more trees to the ensemble", | |
| "Set booster to 'gblinear'" | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." | |
| }, | |
| { | |
| "id": 87, | |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q87)", | |
| "options": [ | |
| "Increase subsample and colsample_bytree values", | |
| "Reduce regularization parameters lambda and alpha", | |
| "Add more trees to the ensemble", | |
| "Set booster to 'gblinear'" | |
| ], | |
| "correctAnswerIndex": 0, | |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." | |
| }, | |
| { | |
| "id": 88, | |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q88)", | |
| "options": [ | |
| "Increase subsample and colsample_bytree values", | |
| "Reduce regularization parameters lambda and alpha", | |
| "Add more trees to the ensemble", | |
| "Set booster to 'gblinear'" | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." | |
| }, | |
| { | |
| "id": 89, | |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q89)", | |
| "options": [ | |
| "Increase subsample and colsample_bytree values", | |
| "Reduce regularization parameters lambda and alpha", | |
| "Add more trees to the ensemble", | |
| "Set booster to 'gblinear'" | |
| ], | |
| "correctAnswerIndex": 1, | |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." | |
| }, | |
| { | |
| "id": 90, | |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q90)", | |
| "options": [ | |
| "Increase subsample and colsample_bytree values", | |
| "Reduce regularization parameters lambda and alpha", | |
| "Add more trees to the ensemble", | |
| "Set booster to 'gblinear'" | |
| ], | |
| "correctAnswerIndex": 0, | |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." | |
| }, | |
| { | |
| "id": 91, | |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q91)", | |
| "options": [ | |
| "Increase subsample and colsample_bytree values", | |
| "Reduce regularization parameters lambda and alpha", | |
| "Add more trees to the ensemble", | |
| "Set booster to 'gblinear'" | |
| ], | |
| "correctAnswerIndex": 1, | |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." | |
| }, | |
| { | |
| "id": 92, | |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q92)", | |
| "options": [ | |
| "Increase subsample and colsample_bytree values", | |
| "Reduce regularization parameters lambda and alpha", | |
| "Add more trees to the ensemble", | |
| "Set booster to 'gblinear'" | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." | |
| }, | |
| { | |
| "id": 93, | |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q93)", | |
| "options": [ | |
| "Increase subsample and colsample_bytree values", | |
| "Reduce regularization parameters lambda and alpha", | |
| "Add more trees to the ensemble", | |
| "Set booster to 'gblinear'" | |
| ], | |
| "correctAnswerIndex": 0, | |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." | |
| }, | |
| { | |
| "id": 94, | |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q94)", | |
| "options": [ | |
| "Increase subsample and colsample_bytree values", | |
| "Reduce regularization parameters lambda and alpha", | |
| "Add more trees to the ensemble", | |
| "Set booster to 'gblinear'" | |
| ], | |
| "correctAnswerIndex": 1, | |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." | |
| }, | |
| { | |
| "id": 95, | |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q95)", | |
| "options": [ | |
| "Increase subsample and colsample_bytree values", | |
| "Reduce regularization parameters lambda and alpha", | |
| "Add more trees to the ensemble", | |
| "Set booster to 'gblinear'" | |
| ], | |
| "correctAnswerIndex": 1, | |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." | |
| }, | |
| { | |
| "id": 96, | |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q96)", | |
| "options": [ | |
| "Increase subsample and colsample_bytree values", | |
| "Reduce regularization parameters lambda and alpha", | |
| "Add more trees to the ensemble", | |
| "Set booster to 'gblinear'" | |
| ], | |
| "correctAnswerIndex": 0, | |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." | |
| }, | |
| { | |
| "id": 97, | |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q97)", | |
| "options": [ | |
| "Increase subsample and colsample_bytree values", | |
| "Reduce regularization parameters lambda and alpha", | |
| "Add more trees to the ensemble", | |
| "Set booster to 'gblinear'" | |
| ], | |
| "correctAnswerIndex": 3, | |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." | |
| }, | |
| { | |
| "id": 98, | |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q98)", | |
| "options": [ | |
| "Increase subsample and colsample_bytree values", | |
| "Reduce regularization parameters lambda and alpha", | |
| "Add more trees to the ensemble", | |
| "Set booster to 'gblinear'" | |
| ], | |
| "correctAnswerIndex": 1, | |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." | |
| }, | |
| { | |
| "id": 99, | |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q99)", | |
| "options": [ | |
| "Increase subsample and colsample_bytree values", | |
| "Reduce regularization parameters lambda and alpha", | |
| "Add more trees to the ensemble", | |
| "Set booster to 'gblinear'" | |
| ], | |
| "correctAnswerIndex": 2, | |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." | |
| }, | |
| { | |
| "id": 100, | |
| "questionText": "Scenario: A data scientist is using XGBoost Regression to predict house prices. She observes the model overfits despite tuning 'max_depth' and 'learning_rate'. Which additional parameter might help reduce overfitting (Hard Q100)", | |
| "options": [ | |
| "Increase subsample and colsample_bytree values", | |
| "Reduce regularization parameters lambda and alpha", | |
| "Add more trees to the ensemble", | |
| "Set booster to 'gblinear'" | |
| ], | |
| "correctAnswerIndex": 1, | |
| "explanation": "Increasing subsample and colsample_bytree introduces randomness by sampling data and features for each tree, helping prevent overfitting." | |
| } | |
| ] | |
| } |