{"id":111100,"date":"2023-07-28T20:55:15","date_gmt":"2023-07-28T20:55:15","guid":{"rendered":"https:\/\/learnexams.com\/blog\/?p=111100"},"modified":"2023-07-28T20:55:22","modified_gmt":"2023-07-28T20:55:22","slug":"isye-6414-final-exam-real-exam-questions-and-answers-2022-2024-graded-a-exam-1","status":"publish","type":"post","link":"https:\/\/www.learnexams.com\/blog\/2023\/07\/28\/isye-6414-final-exam-real-exam-questions-and-answers-2022-2024-graded-a-exam-1\/","title":{"rendered":"ISYE 6414 FINAL EXAM (REAL EXAM) QUESTIONS AND ANSWERS 2022-2024\/ GRADED A | EXAM 1"},"content":{"rendered":"\n<p>True &#8211; The relationship that links the predictors is highly non-linear.<br>In Logistic Regression, the relationship between the probability of success and the predicting variables is non-linear.<\/p>\n\n\n\n<p>False &#8211; In logistic regression, there are no error terms.<br>In Logistic Regression, the error terms follow a normal distribution.<\/p>\n\n\n\n<p>True &#8211; the logit function is also known as the log-odds function, which is the ln(P\/1-p).<br>The logit function is the log of the ratio of the probability of success to the probability of failure and is also known as the log-odds function.<\/p>\n\n\n\n<p>False &#8211; As there is no error term in logistic regression, there is no additional parameter for the variance of the error terms.<br>The number of parameters that need to be estimated in a logistic regression model with 6 predicting variables and an intercept is the same as the number of parameters that need to be estimated in a standard linear regression model with an intercept and same predicting variables.<\/p>\n\n\n\n<p>False &#8211; log-likelihood is a non-linear function, and a numerical algorithm is needed in order to maximize it.<br>The log-likelihood function is a linear function with a closed form solution.<\/p>\n\n\n\n<p>False &#8211; We interpret logistic regression coefficients with respect to the odds of success.<br>In Logistic Regression, the estimated value for a regression coefficient B represents the estimated expected change in the response variable associated with a one unit increase in the predicting variable, holding all else fixed.<\/p>\n\n\n\n<p>False &#8211; The coefficient estimator follows an approximate normal distribution.<br>Under logistic regression, the sampling distribution used for a coefficient estimator is a chi-square distribution when the sample size is large.<\/p>\n\n\n\n<p>False &#8211; when testing a subset of coefficients, deviance follows a chi-square distribution with q degrees of freedom, where q is the number of regression coefficients discarded from the full model to get the reduced model.<br>When testing a subset of coefficients, deviance follows a chi-square distribution with q degrees of freedom, where q is the number of regression coefficients in the reduced model.<\/p>\n\n\n\n<p>True &#8211; logistic regression is the generalization of the standard regression model that is used when the response variable y is binary or binomial.<br>Logistic regression deals with the case where the dependent variable is binary and the conditional distribution is binomial.<\/p>\n\n\n\n<p>False &#8211; The residuals can only be defined for logistic regression with replications.<br>It is good practice to perform a goodness-of-fit test on logistic regression models without replications.<\/p>\n\n\n\n<p>False &#8211; for logistic regression, if the p-value of the deviance test for GOD is large, then the model is a good fit.<br>In Logistic regression, if the p-value of the deviance test for GOF is smaller than the significance level alpha, then is is plausible that the model is a good fit.<\/p>\n\n\n\n<p>False &#8211; GOF is no guarantee for good prediction and vice-versa.<br>If a logistic regression model provides accurate classification, then we can conclude that it is a good fir for the data.<\/p>\n\n\n\n<p>True &#8211; the deviance residuals are approximately N(0,1) if the model is a good fit to the data.<br>For both logistic regression and Poisson regression, the deviance residuals should follow an approximate standard normal distribution if the model is a good fit for the data.<\/p>\n\n\n\n<p>False<br>The logit link function is the best link function to model binary response data because it always fits the data better than other link functions.<\/p>\n\n\n\n<p>True &#8211; we can use the Pearson or deviance residuals, but only if the model has replications.<br>Although there are no error terms in logistic regression model using binary data with replications, we can still perform residual analysis.<\/p>\n\n\n\n<p>True &#8211; The error rate is biased downwards, since the model sees the data 2 times, once for training and once for testing.<br>For a classification model, the training error tends to underestimate the true classification error rate of the model.<\/p>\n\n\n\n<p>True &#8211; the parameters and their standard errors are approximate.<br>The estimated regression coefficients in Poisson regression are approximate.<\/p>\n\n\n\n<p>False &#8211; we use a z-test, since the the distributions are approximately normal with large N.<br>A t-test is used for testing the statistical significance of a coefficient given all predicting variables in a Poisson regression model.<\/p>\n\n\n\n<p>True<br>An overdispersion parameter of 1 indicates that the variability of the response is close to the variability estimated by the model.<\/p>\n\n\n\n<p>False &#8211; we assume that the log rate is a linear combination of the predicting variables, hence Poisson regression is a generalized linear model (GLM)<br>In Poisson regression, we assume a non linear relationship between the log rate and the predicting variables.<\/p>\n\n\n\n<p>True<br>Logistic regression models the probability of a success given a set of predicting variables.<\/p>\n\n\n\n<p>True<br>The estimation of logistic regression coefficients is based on maximum likelihood.<\/p>\n\n\n\n<p>1) no error term<br>2) the response variable is not normally distributes (binomial)<br>3) it models probability, not expectation of response<br>What are the differences between logistic regression and standard regression.<\/p>\n\n\n\n<p>False &#8211; Come back and name other link functions<br>The logit link function is the only link function that can be used for modeling binary response data.<\/p>\n\n\n\n<p>False &#8211; logistic regression coefficients are interpreted with respect to odds<br>The interpretation of the regression coefficients is the same in logistic regression as standard regression.<\/p>\n\n\n\n<p>False &#8211; there is no closed form solution, we we use a numerical approximation.<br>We can derive exact estimates for the logistic regression coefficients.<\/p>\n\n\n\n<p>False &#8211; fill in later<br>The estimations of the regression coefficients is based on minimizing the sum of least squares in logistic regression.<\/p>\n\n\n\n<p>1) The sampling distribution of the regression coefficients is approximate<br>2) a large sample size is required for making accurate statistical inferences<br>3) a normal sampling distribution is used instead of a t-distribution for statistical inference<br>Differences between logistic regression and linear regression &#8211; statistical inference.<\/p>\n\n\n\n<p>True &#8211; statistical inference in logistic regression is only reliable when N is large<br>in logistic regression, the hypothesis test for subsets of coefficients is approximate, it relies on a large sample size.<\/p>\n\n\n\n<p>True &#8211; we predict if a response will be a success or failure<br>In logistic regression, prediction is a classification of a future binary response.<\/p>\n\n\n\n<p>True<br>in k-fold cross validation, the larger K, the higher the variability in the estimation of the classification error is.<\/p>\n\n\n\n<p>1) to model count data<br>2) to model rate response data<br>3) to model response data with a Poisson distribution<br>What can Poisson regression be used for?<\/p>\n\n\n\n<p>True<br>The link function for the Poisson regression is the log function.<\/p>\n\n\n\n<p>False &#8211; constant variance will be violated.<br>If we apply a standard regression to response data with a Poisson distribution, constant variance assumption will hold.<\/p>\n\n\n\n<p>True<br>in Poisson regression, we model the log of the expected response variable, not the expected log response variable.<\/p>\n\n\n\n<p>False &#8211; fill in later<br>In Poisson regression, we use ordinary least squares to fit the model.<\/p>\n\n\n\n<p>True<br>In Poisson regression, we interpret the coefficients in terms of the ratio of the response rates.<\/p>\n\n\n\n<p>False &#8211; we use z-tests<br>In Poisson regression, we make inference using the t-intervals for the coefficients.<\/p>\n\n\n\n<p>False &#8211; the estimates for the coefficients are approximate in Poisson regression.<br>in Poisson regression, inference relies on the exact sampling distribution of the regression coefficients.<\/p>\n\n\n\n<p>True &#8211; the test for regression coefficients in Poisson regression follows a chi-square distribution with q degrees of freedom.<br>We use a chi-square testing procedure to test whether a subset of regression coefficients are zero in Poisson regression.<\/p>\n\n\n\n<p>False &#8211; fill in later<br>We can use residual analysis in Poisson regression to evaluate whether errors are uncorrelated.<\/p>\n\n\n\n<p>1) to address multicollinearity in multiple regression<br>2) To select among a large number of predicting variables<br>3) To fit a model when there are more predicting variables than observations<br>What are some common use cases for variable selection?<\/p>\n\n\n\n<p>True<br>When selecting variables, it is important to first establish which variables are used for controlling bias in the sample and which are explanatory.<\/p>\n\n\n\n<p>True -Variable selection balances bias with variance to select the model.<br>Variable selection methods are performed by balancing the bias-variance tradeoff.<\/p>\n\n\n\n<p>True<br>The penalty constant Lambda in regularized regression has the role of controlling the trade off between lack of fit and model complexity.<\/p>\n\n\n\n<p>True &#8211; we can find closed form solutions for the ridge coefficients<br>The ridge regression coefficients are obtained using an exact or closed form expression.<\/p>\n\n\n\n<p>True &#8211; in Lasso, the coefficient estimates are approximate, we used a numerical algorithm to estimate them.<br>The estimated coefficients in lasso regression are obtained using a numerical algorithm.<\/p>\n\n\n\n<p>True &#8211; fill in later<br>The regression coefficients in lasso are less efficient than those from the ordinary least squares estimation approach.<\/p>\n\n\n\n<p>True &#8211; this is true for explanatory purposes but NOT prediction.<br>When Selecting variables for explanatory purpose, one might consider including predicting variables which are correlated if it would help answer your research hypothesis.<\/p>\n\n\n\n<p>False &#8211; Variable selection has come a long way but is far from a solved problem, especially with many predictors.<br>Variable selection is a simple and solved statistical problem since we can implement it using software.<\/p>\n\n\n\n<p>False &#8211; it is not good practice to perform variable selection based on the statistical significance of the coefficients, as significance is almost always derived based on the other predictors in the model.<br>It is good practice to perform variable selection based on the statistical significance of the regression coefficients.<\/p>\n\n\n\n<p>True &#8211; since the model sees the data 2 times, training risk is generally too favorable when estimating true prediction risk.<br>The training risk is a biased estimator for prediction risk.<\/p>\n\n\n\n<p>True &#8211; AIC is a measure for prediction risk that adds a penalty term to correct for the bias in the training risk.<br>AIC is an estimate for the prediction risk.<\/p>\n\n\n\n<p>True &#8211; BIC generally penalized more than all of the other prediction risk estimated, and is especially useful when the objective is prediction, since it selects simpler models.<br>BIC penalizes for complexity of the model more than both VC and Mallow&#8217;s Cp statistic.<\/p>\n\n\n\n<p>False &#8211; often times these 2 approaches will not select the same set of variables.<br>When the number of predicting variables is large, both backward and forward stepwise regression will always select the same set of variables.<\/p>\n\n\n\n<p>False &#8211; complex models with many predictors have low bias but high variance.<br>Complex models with many predictors are often extremely biased, but have low variance.<\/p>\n\n\n\n<p>False &#8211; Backward stepwise regression is more computationally expensive than forward and generally selects a larger model. Forward is usually preferred over backward.<br>Backward stepwise regression is preferable over forward stepwise regression because it starts with larger models.<\/p>\n\n\n\n<p>False &#8211; stepwise regression is greedy, which means that it is a heuristic and not all combinations of the model are checked.<br>Stepwise regression is a greedy algorithm searching through all possible combinations of the predicting variables to find the model with the best score.<\/p>\n\n\n\n<p>True &#8211; if some variables are required for whatever reason, we should force the minimum model to add those variables.<br>If specific variables need to be included to control bias in the model, they should be forced into the model and not be a part of the variable selection process.<\/p>\n\n\n\n<p>False &#8211; the L1 penalty measures sparsity and is associated with Lasso, which forces coefficients to 0. The L2 penalty is used to deal with multicollinearity and does not force coefficients to 0.<br>The L2 penalty measures the sparsity of a vector and forces regression coefficients to be zero.<\/p>\n\n\n\n<p>True &#8211; Elastic net regression combines the L1 and L2 benefits, but it also combines the disadvantages of both approaches!<br>Elastic net regression uses both penalties of ridge and lasso regression and hence combined the benefits of both.<\/p>\n\n\n\n<p>False &#8211; regularized regression requires the predictors to be scaled.<br>It is not requires to standardize or rescale the predicting variables when performing regularized regression.<\/p>\n\n\n\n<p>False &#8211; Ridge regression is associated with the L2 penalty, which does not perform variable selection.<br>Ridge Regression is a regularized regression approach that can be used for variable selection.<\/p>\n\n\n\n<p>True &#8211; there is no closed-form solution for lasso regression coefficient estimates.<br>The lasso regression requires a numerical algorithm to minimize the penalized sum of least squares.<\/p>\n\n\n\n<p>False &#8211; the shrinkage penalty is applied to all b1\u2026BP, but not the intercept B0.<br>In regularized regression, the penalization is generally applied to all regression coefficients, where p = number of predictors.<\/p>\n\n\n\n<p>True &#8211; Lambda balances the bias-variance tradeoff.<br>The penalty constant Lambda is penalized regression controls the trade off between lack of fit and model complexity.<\/p>\n\n\n\n<p>False &#8211; when Lambda = 0, the corresponding regression coefficients are equal to OLS.<br>In ridge regression, when the penalty constant Lambda = 1, the corresponding ridge regression coefficients are the same as OLS.<\/p>\n\n\n\n<p>False &#8211; Ridge regression used the L2 penalty, which specifically addresses multicollinearity in the predictors.<br>Ridge Regression cannot be used to deal with problems caused by high correlation among predictors.<\/p>\n\n\n\n<p>True<br>We can obtain both the estimates and the standard deviations of the estimates for the regression coefficients in logistic regression.<\/p>\n\n\n\n<p>True<br>In Logistic regression, the sampling distribution of the residual is approximately normal if the data is a good fit.<\/p>\n\n\n\n<p>False<br>In logistic regression, the residuals are derived as the fitted values minus the observed responses.<\/p>\n\n\n\n<p>Least Square Elimination (LSE) cannot be applied to GLM models.<br>False &#8211; it is applicable but does not use data distribution information fully.<\/p>\n\n\n\n<p>In multiple linear regression with idd and equal variance, the least squares estimation of regression coefficients are always unbiased.<br>True &#8211; the least squares estimates are BLUE (Best Linear Unbiased Estimates) in multiple linear regression.<\/p>\n\n\n\n<p>Maximum Likelihood Estimation is not applicable for simple linear regression and multiple linear regression.<br>False &#8211; In SLR and MLR, the SLE and MLE are the same with normal idd data.<\/p>\n\n\n\n<p>The backward elimination requires a pre-set probability of type II error<br>False &#8211; Type I error<\/p>\n\n\n\n<p>The first degree of freedom in the F distribution for any of the three procedures in stepwise is always equal to one.<br>True<\/p>\n\n\n\n<p>MLE is used for the GLMs for handling complicated link function modeling in the X-Y relationship.<br>True<\/p>\n\n\n\n<p>In the GLMs the link function cannot be a non linear regression.<br>False &#8211; It can be linear, non linear, or parametric<\/p>\n\n\n\n<p>When the p-value of the slope estimate in the SLR is small the r-squared becomes smaller too.<br>False &#8211; When P value is small, the model fits become more significant and R squared become larger.<\/p>\n\n\n\n<p>In GLMs the main reason one does not use LSE to estimate model parameters is the potential constrained in the parameters.<br>False &#8211; The potential constraint in the parameters of GLMs is handled by the link function.<\/p>\n\n\n\n<p>The R-squared and adjusted R-squared are not appropriate model comparisons for non linear regression but are for linear regression models.<br>TRUE &#8211; The underlying assumption of R-squared calculations is that you are fitting a linear model.<\/p>\n\n\n\n<p>The decision in using ANOVA table for testing whether a model is significant depends on the normal distribution of the response variable<br>True<\/p>\n\n\n\n<p>When the data may not be normally distributed, AIC is more appropriate for variable selection than adjusted R-squared<br>True<\/p>\n\n\n\n<p>The slope of a linear regression equation is an example of a correlation coefficient.<br>False &#8211; the correlation coefficient is the r value. Will have the same + or &#8211; sign as the slope.<\/p>\n\n\n\n<p>In multiple linear regression, as the value of R-squared increases, the relationship<br>between predictors becomes stronger<br>False &#8211; r squared measures how much variability is explained by the model, NOT how strong the predictors are.<\/p>\n\n\n\n<p>When dealing with a multiple linear regression model, an adjusted R-squared can<br>be greater than the corresponding unadjusted R-Squared value.<br>False &#8211; the adjusted rsquared value take the number and types of predictors into account. It is lower than the r squared value.<\/p>\n\n\n\n<p>In a multiple regression problem, a quantitative input variable x is replaced by x \u2212<br>mean(x). The R-squared for the fitted model will be the same<br>True<\/p>\n\n\n\n<p>The estimated coefficients of a regression line is positive, when the coefficient of<br>determination is positive.<br>False &#8211; r squared is always positive.<\/p>\n\n\n\n<p>If the outcome variable is quantitative and all explanatory variables take values 0 or<br>1, a logistic regression model is most appropriate.<br>False &#8211; More research is necessary to determine the correct model.<\/p>\n\n\n\n<p>After fitting a logistic regression model, a plot of residuals versus fitted values is<br>useful for checking if model assumptions are violated.<br>False &#8211; for logistic regression use deviance residuals.<\/p>\n\n\n\n<p>In a greenhouse experiment with several predictors, the response variable is the<br>number of seeds that germinate out of 60 that are planted with different treatment<br>combinations. A Poisson regression model is most appropriate for modeling this<br>data<br>False &#8211; poisson regression models rate or count data.<\/p>\n\n\n\n<p>For Poisson regression, we can reduce type I errors of identifying statistical<br>significance in the regression coefficients by increasing the sample size.<br>True<\/p>\n\n\n\n<p>Both LASSO and ridge regression always provide greater residual sum of squares<br>than that of simple multiple linear regression.<br>True<\/p>\n\n\n\n<p>If data on (Y, X) are available at only two values of X, then the model Y = \\beta_1 X<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>\\beta_2 X^2 + \\epsilon provides a better fit than Y = \\beta_0 + \\beta_1 X +<br>\\epsilon.<br>False &#8211; nothing to determine of a quadratic model is necessary or required.<\/li>\n<\/ul>\n\n\n\n<p>If the Cook&#8217;s distance for any particular observation is greater than one, that data<br>point is definitely a record error and thus needs to be discarded.<br>False &#8211; must see a comparison of data points. Is 1 too large?<\/p>\n\n\n\n<p>We can use residual analysis to conclusively determine the assumption of<br>independence<br>False &#8211; we can only determine uncorrelated errors.<\/p>\n\n\n\n<p>It is possible to apply logistic regression when the response variable Y has 3<br>classes.<br>True<\/p>\n\n\n\n<p>. A correlation coefficient close to 1 is evidence of a cause-and-effect relationship<br>between the two variables.<br>False- cause and effect can only be determined by a well designed experiment.<\/p>\n\n\n\n<p>Multiplying a variable by 10 in LASSO regression, decreases the chance that the<br>coefficient of this variable is nonzero.<br>False &#8211; I am not sure why anyone would think this would be true.<\/p>\n\n\n\n<p>In regression inference, the 99% confidence interval of coefficient \\beta_0 is always<br>wider than the 95% confidence interval of \\beta_1.<br>False- can only compare beta1 with beta1 and beta0 with beta0<\/p>\n\n\n\n<p>The regression coefficients for the Poisson regression model can be estimated in<br>exact\/closed form.<br>False &#8211; MLE is NOT closed form.<\/p>\n\n\n\n<p>Mean square error is commonly used in statistics to obtain estimators that may be biased, but less uncertain than unbiased ones. And that&#8217;s preferred.<br>True<\/p>\n\n\n\n<p>Regression models are only appropriate for continuous response variables.<br>False &#8211; logistic and poisson model probability and rate<\/p>\n\n\n\n<p>The assumptions in logistic regression are &#8211; Linearity, Independence of response variable, and the link function is the logit function.<br>True &#8211; linearity is measured through the link, , the g of the probability of success and the predicted variable.<\/p>\n\n\n\n<p>The log odds function, also called the logit function, which is the log of the ratio between the probability of a success and the probability of a failure<br>True<\/p>\n\n\n\n<p>In logistic regression we interpret the Betas in terms of the response variable.<br>False &#8211; we interpret it in terms of the odds of success or the log odds of success<\/p>\n\n\n\n<p>In logistic regression we have an additional error term to estimate.<br>False &#8211; there is not error term in logistic regression.<\/p>\n\n\n\n<p>The least square estimation for the standard regression model is equivalent with Maximum Likelihood Estimation, under the assumption of normality.<br>True<\/p>\n\n\n\n<p>The variance estimator in logistic regression has a closed form expression.<br>False &#8211; use statistical software to obtain the variance-co-variance matrix<\/p>\n\n\n\n<p>We can use the z value to determine if a coefficient is equal to zero in logistic regression.<br>True &#8211; z value = (Beta-0)\/(SE of Beta)<\/p>\n\n\n\n<p>In testing for a subset of coefficients in logistic regression the null hypothesis is that the coefficient is equal to zero<br>True<\/p>\n\n\n\n<p>Like standard linear regression we can use the F test to test for overall regression in logistic regression.<br>False &#8211; It&#8217;s 1-pchisq(null deviance-residual deviance, DFnull-DFresidual)<\/p>\n\n\n\n<p>For logistic regression we can define residuals for evaluating model goodness of fit for models with and without replication.<br>False &#8211; can only be with replication under the assumption that Yi is binary and n1 is greater than 1<\/p>\n\n\n\n<p>The deviance residuals are the signed square root of the log-likelihood evaluated at the saturated model<br>True<\/p>\n\n\n\n<p>From the binomial approximation with a normal distribution using the central limit theorem, the Pearson residuals have an approximately standard chi-squared distribution.<br>False &#8211; Normal distribution<\/p>\n\n\n\n<p>Visual Analytics for logistic regression<br>Normal probability plot of residuals<br>Residuals vs predictors<br>Logit of success rate vs predictors<br>True<br>Normal probability plot of residuals &#8211; Normality<br>Residuals vs predictors &#8211; Linearity\/Independence<br>Logit of success rate vs predictors &#8211; Linearity<\/p>\n\n\n\n<p>Under the null hypothesis of good fit for logistic regression, the test statistic has a Chi-Square distribution with n- p- 1 degrees of freedom<br>True &#8211; don&#8217;t forget, we want large P values<\/p>\n\n\n\n<p>For the testing procedure for subsets of coefficients, we compare the likelihood of a reduced model versus a full model. This is a goodness of fit test<br>False &#8211; it provides inference of the predictive power of the model<\/p>\n\n\n\n<p>Predictive power means that the predicting variables predict the data even if one or more of the assumptions do not hold.<br>True<\/p>\n\n\n\n<p>One reason why the logistic model may not fit is the relationship between logit of the expected probability and predictors might be multiplicative, rather than additive<br>True<\/p>\n\n\n\n<p>In logistic regression for goodness of fit, we can only use the Pearson residuals.<br>False &#8211; we can use Pearson or Deviance.<\/p>\n\n\n\n<p>An indication that a higher order non linear relationship better fits the data is that the dummy variables are all, or nearly all, statistically significant<br>True<\/p>\n\n\n\n<p>Simpson&#8217;s Paradox &#8211; the reversal of association when looking at marginal vs conditional relationships<br>True<\/p>\n\n\n\n<p>Classification is nothing else than prediction of binary responses.<br>True<\/p>\n\n\n\n<p>We cannot use the training error rate as an estimate of the true error classification error rate because it is biased upward.<br>False &#8211; biased downward<\/p>\n\n\n\n<p>Random sampling is computationally more expensive than the K-fold cross validation, with no clear advantage in terms of the accuracy of the estimation classification error rate.<br>True<\/p>\n\n\n\n<p>Leave on out cross validation is preferred<br>False &#8211; K fold is preferred.<\/p>\n\n\n\n<p>The larger K is, the larger the number of folds, the less bias the estimate of the classification the error is but has higher variability.<br>True<\/p>\n\n\n\n<p>In Poisson regression underlying assumption is that the response variable has a Poisson distribution, or responses could be wait times, or exponential distribution<br>True<\/p>\n\n\n\n<p>The g link function is also called the canonical link function.<br>True &#8211; which means that parameter estimates under logistic regression are fully efficient and tests on those parameters are better behaved for small samples.<\/p>\n\n\n\n<p>Poisson distribution, the variance is equal to the expectation. Thus, the variance is not constant<br>True<\/p>\n\n\n\n<p>For Poisson regression we estimate the expectation of the log response variable.<br>False &#8211; we estimate the log of the expectation of the response variable.<\/p>\n\n\n\n<p>Standard linear regression could be used to model Poisson regression using the variance stabilizing transformation sqrt(mu-3\/8) if the number of counts is large<br>True &#8211; the number of counts can be small &#8211; then use Poisson<\/p>\n\n\n\n<p>In Poisson Regression we do not interpret beta with respect to the response variable but with respect to the ratio of the rate.<br>True<\/p>\n\n\n\n<p>In Poisson regression we model the error term<br>False &#8211; there is no error term<\/p>\n\n\n\n<p>One problem with fitting a normal regression model to Poisson data is the departure from the assumption of constant variance<br>True<\/p>\n\n\n\n<p>Event rates can be calculated as events per units of varying size, this unit of size is called exposure<br>True<\/p>\n\n\n\n<p>The estimators for the regression coefficients in the Poisson regression are biased.<br>False &#8211; they are unbiased<\/p>\n\n\n\n<p>To perform hypothesis testing for Poisson, we can use again the approximate normal sampling distribution, also called the Wald test<br>True &#8211; Wald Test also used with logistic regression<\/p>\n\n\n\n<p>Hypothesis testing for Poisson regression can be done on small sample sizes<br>False &#8211; Approximation of normal distribution needs large sample sizes, so does hypothesis testing.<\/p>\n\n\n\n<p>For large sample size data, the distribution of the test statistic, assuming the null hypothesis, is a chi-squared distribution<br>True<\/p>\n\n\n\n<p>The p-value of the test computed as the left tail of the chi-squared distribution<br>False &#8211; Right tail<\/p>\n\n\n\n<p>Poisson Assumptions &#8211; log transformation of the rate is a linear combination of the predicting variables, the response variables are independently observed, the link function g is the log function<br>True &#8211; remember, NO ERROR TERM<\/p>\n\n\n\n<p>Overdispersion is when the variability of the response variable is larger than estimated by the model<br>True<\/p>\n\n\n\n<p>The gam() function is a non-parametric test to determine what transformation is best.<br>True<\/p>\n\n\n\n<p>The deviance and pearson residuals are normally distributed<br>TRUE &#8211; the residual deviances are chi square distributed<\/p>\n\n\n\n<p>Model with many predictors have high bias but low variance.<br>False &#8211; low bias and high variance<\/p>\n\n\n\n<p>When the objective is to explain the relationship to the response, one might consider including predicting variables which are correlated<br>True &#8211; But this should be avoided for prediction<\/p>\n\n\n\n<p>Variable selection addresses multicolinearity, high dimensionaltiy, and prediction vs explanatory prediction<br>TRUE<\/p>\n\n\n\n<p>The variables chosen for prediction and the variables chosen for explanatory objectives will be the same.<br>False<\/p>\n\n\n\n<p>Variable selection is not special, it is affected by highly correlated variables<br>TRUE<\/p>\n\n\n\n<p>Confounding variable is a variable that influences both the dependent variable and independent variable<br>True<\/p>\n\n\n\n<p>Explanatory variable is one that explains changes in the response variable<br>TRUE<\/p>\n\n\n\n<p>Predicting variable is used in regression to predict the outcome of another variable.<br>True<\/p>\n\n\n\n<p>It is good practice apply variable selection without understanding the problem at hand to reduce bias.<br>False &#8211; always understand the problem at hand to better select variables for the model.<\/p>\n\n\n\n<p>When a statistically insignificant variable is discarded from the model, there is little change in the other predictors statistical significance.<br>False &#8211; it is possible that when a predictor is discarded, the statistical significance of other variables will change.<\/p>\n\n\n\n<p>We can do a partial F test to determine if variable selection is necessary.<br>True<\/p>\n\n\n\n<p>When selecting variables for a model, one needs also to consider the research hypothesis, as well as any potential confounding variables to control for<br>True<\/p>\n\n\n\n<p>We would like to have a prediction with low uncertainty for new settings. This means that we&#8217;re willing to give up some bias to reduce the variability in the prediction.<br>True<\/p>\n\n\n\n<p>Generally models with covariance have high bias but low variance<br>False &#8211; they have low bias but high variance.<\/p>\n\n\n\n<p>A measure of the bias-variance tradeoff is the prediction risk<br>TRUE<\/p>\n\n\n\n<p>To estimate prediction risk we compute the prediction risk for the observed data and take the sum of squared differences between fitted values for sub model S and the observed values.<br>True &#8211; this is called training risk and it is a biased estimate of prediction risk<\/p>\n\n\n\n<p>The larger the number of variables in the model, the larger the training risk.<br>False &#8211; the larger the number of variables in a model the lower the training risk.<\/p>\n\n\n\n<p>The Mallow&#8217;s CP complexity penalty is two times the size of the model (the number of variables in the submodel) times the estimated variance divided by n.<br>True<\/p>\n\n\n\n<p>AIC looks just like the Mallow&#8217;s Cp except that the variance is the true variance and not its estimate.<br>True<\/p>\n\n\n\n<p>Another criteria for variable selection is cross validation which is a direct measure of explanatory power.<br>False &#8211; Predictive power<\/p>\n\n\n\n<p>Stepwise is a heuristic search<br>TRUE it is also a greedy search that does not guarantee to find the best score<\/p>\n\n\n\n<p>If p is larger than n, stepwise is feasible<br>TRUE &#8211; for forward, but not backward<\/p>\n\n\n\n<p>Forward stepwise will select larger models than backward.<br>False &#8211; it will typically select smaller models especially if p is large<\/p>\n\n\n\n<p>Mallow&#8217;s CP is useful when there are no control variables.<br>TRUE<\/p>\n\n\n\n<p>The overall regression F-statistic tests the null hypothesis that<br>the coefficients are equal to zero<\/p>\n\n\n\n<p>The test of subset of coefficients tests the null hypothesis that<br>discarded variables have coefficients equal to zero.<\/p>\n\n\n\n<p>Goodness of fit tests the null hypothesis that<br>the model fits the data<\/p>\n\n\n\n<p>The prediction risk is the sum between the irreducible error and the mean square error<br>True<\/p>\n\n\n\n<p>There is never a situation where a complex model is best.<br>False &#8211; there are situations where a complex model is best<\/p>\n\n\n\n<p>L0 penalty, which is the number of nonzero regression coefficients<br>True &#8211; not feasible for a large number of predicting variables as requires fitting all models<\/p>\n\n\n\n<p>L1 penalty will force many betas, many regression coefficients to be 0s<br>True &#8211; is equal to the sum of the absolute values of the regression coefficients to be penalized<\/p>\n\n\n\n<p>L2 does not perform variable selection<br>True &#8211; is equal to the sum of the squared regression coefficients to be penalized and does not do variable selection<\/p>\n\n\n\n<p>L2 penalty term measures sparsity<br>False &#8211; L1 penalty measures sparsity. L2 removes the limitation on variable selection<\/p>\n\n\n\n<p>The estimated regression coefficients from Lasso are less efficient than those provided by the ordinary least squares<br>True<\/p>\n\n\n\n<p>Where p the number of predictors is larger than n the number of observationsthe Lasso selects, at most, n variables<br>True when p is greater than n, lasso will select n variables at the most<\/p>\n\n\n\n<p>If there is a high correlation between variables, Lasso will select both.<br>False lasso will select 1<\/p>\n\n\n\n<p>Isye 6414 final exam quizlet<br>Isye 6414 final exam questions<br>Isye 6414 final exam answers<br>isye 6414 final coursehero<br>isye 6414 final exam cheat sheet<br>isye 6414 midterm 1<\/p>\n","protected":false},"excerpt":{"rendered":"<p>True &#8211; The relationship that links the predictors is highly non-linear.In Logistic Regression, the relationship between the probability of success and the predicting variables is non-linear. False &#8211; In logistic regression, there are no error terms.In Logistic Regression, the error terms follow a normal distribution. True &#8211; the logit function is also known as the [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"site-sidebar-layout":"default","site-content-layout":"","ast-site-content-layout":"default","site-content-style":"default","site-sidebar-style":"default","ast-global-header-display":"","ast-banner-title-visibility":"","ast-main-header-display":"","ast-hfb-above-header-display":"","ast-hfb-below-header-display":"","ast-hfb-mobile-header-display":"","site-post-title":"","ast-breadcrumbs-content":"","ast-featured-img":"","footer-sml-layout":"","ast-disable-related-posts":"","theme-transparent-header-meta":"","adv-header-id-meta":"","stick-header-meta":"","header-above-stick-meta":"","header-main-stick-meta":"","header-below-stick-meta":"","astra-migrate-meta-layouts":"default","ast-page-background-enabled":"default","ast-page-background-meta":{"desktop":{"background-color":"","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"tablet":{"background-color":"","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"mobile":{"background-color":"","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""}},"ast-content-background-meta":{"desktop":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"tablet":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"mobile":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""}},"footnotes":""},"categories":[],"tags":[],"class_list":["post-111100","post","type-post","status-publish","format-standard","hentry"],"_links":{"self":[{"href":"https:\/\/www.learnexams.com\/blog\/wp-json\/wp\/v2\/posts\/111100","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.learnexams.com\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.learnexams.com\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.learnexams.com\/blog\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.learnexams.com\/blog\/wp-json\/wp\/v2\/comments?post=111100"}],"version-history":[{"count":0,"href":"https:\/\/www.learnexams.com\/blog\/wp-json\/wp\/v2\/posts\/111100\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.learnexams.com\/blog\/wp-json\/wp\/v2\/media?parent=111100"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.learnexams.com\/blog\/wp-json\/wp\/v2\/categories?post=111100"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.learnexams.com\/blog\/wp-json\/wp\/v2\/tags?post=111100"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}