Multinomial Logistic regression Predict in R - r
I am using multinom function from nnet package for multinomial logistic regression. My dataset has 3 features and 14 different classes with total of 1000 observations.
Classes that I have is: 0,1,2,3,4,5,6,7,8,9,10,11,13,15
I divide data set into proper training and calibration, where calibration has only only class of labels(say 4). Training set has all classes except 4.
Now, I train the model as
modelfit <- multinom(label ~ x1+x2+x3, data = train)
Now, I use calibration data to find predicted probabilities as:
predProb = predict(modelfit, newdata=calib_set, type="prob")
where calib_set has only three features and no column of Y.
Then, the predProb gives me the probabilities of all 16 classes except class 11 for all observations in calibration data.
Also, when I use any test data point, I get predicted probabilties of all classes except class 11.
Can, someone explain why is it missing that and how can I get predicted probabilities of all classes?
The below picture shows predicted probabiltiies for calibration data, it is missing class 11, (it can miss class 12 and 14 because that are not in the classes)
Any suggestions or advices are much appreciated.
Related
Difference between fitted values and cross validation values from pls model in r
I only have a small dataset of 30 samples, so I only have a training data set but no test set. So I want to use cross-validation to assess the model. I have run pls models in r using cross-validation and LOO. The mvr output has the fitted values and validation$preds values, and these are different. As the final results of R2 and RMSE for just the training set should I be using the final fitted values or the validation$preds values?
Short answer is if you want to know how good the model is at predicting, you will use the validation$preds because it is tested on unseen data. The values under $fitted.values are obtained by fitting the final model on all your training data, meaning the same training data is used in constructing model and prediction. So values obtained from this final fit, will underestimate the performance of your model on unseen data. You probably need to explain what you mean by "valid" (in your comments). Cross-validation is used to find which is the best hyperparameter, in this case number of components for the model. During cross-validation one part of the data is not used for fitting and serves as a test set. This actually provides a rough estimate the model will work on unseen data. See this image from scikit learn for how CV works. LOO works in a similar way. After finding the best parameter supposedly you obtain a final model to be used on the test set. In this case, mvr trains on all models from 2-6 PCs, but $fitted.values is coming from a model trained on all the training data. You can also see below how different they are, first I fit a model library(pls) library(mlbench) data(BostonHousing) set.seed(1010) idx = sample(nrow(BostonHousing),400) trainData = BostonHousing[idx,] testData = BostonHousing[-idx,] mdl <- mvr(medv ~ ., 4, data = trainData, validation = "CV", method = "oscorespls") Then we calculate mean RMSE in CV, full training model, and test data, using 4 PCs: calc_RMSE = function(pred,actual){ mean((pred - actual)^2)} # error in CV calc_RMSE(mdl$validation$pred[,,4],trainData$medv) [1] 43.98548 # error on full training model , not very useful calc_RMSE(mdl$fitted.values[,,4],trainData$medv) [1] 40.99985 # error on test data calc_RMSE(predict(mdl,testData,ncomp=4),testData$medv) [1] 42.14615 You can see the error on cross-validation is closer to what you get if you have test data. Again this really depends on your data.
Derive preidcted values after glm() in order to get PPV, NPV, sensitivity and specificity
I have run a glm() model; but now I would like to measure the model's accuracy with PPV, NPV, sensitivity and specificity. However, I keep getting confusing results. My outcome is a factor variable that looks like this: table(mydata$outcome) 0 1 6824 359 The predictors are a combination of continuous variables with 1 categorical (gender). My code is: # To run the logistic model mod <- glm(outcome~predictor1+predictor2+predictor3,data=mydata,family=binomial("logit")) summary(mod) # To run predict() to get the predicted values of the outcome predicted = predict(object = mod, newdata=mydata, type = "response") The results for this look like this: head(predicted) 1 2 3 4 5 6 0.02568802 0.02979873 0.01920584 0.01077031 0.01279325 0.09725329 This is very surprising as I was expected to observe predicted '1' (cases) vs '0' (controls) which I could further use to obtain the accuracy measures of the models either with confusionMatrix(predicted, mydata$outcome) or using ModelMetrics library. So my question is how can I get 4x4 table (predicted vs observed) outcome which I can use to measure the accuracy of my glm() model in predicting the outcome? I will be grateful for any advice, or please let me know if there are better ways of getting the PPV, NPV, sensitivity and specificity. Thank you.
Your glm model is giving probabilities of the two outcomes. Typically, one wants to assign '1' to any event with probability >= . 5 and 0 otherwise. You can do this with round(). In more 'machine-learny' type situations, one might consider different values besides .5. You can use the ifelse() fn to do that. For example, if you want to assign '1' only to cases with probability .7 you could say vals = ifelse(mydata$outcome >.7,1,0 ). Finally, the data you want is usually called a confusion matrix. It can be computed via various packages, but here is a nice solution from a sister site - R: how to make a confusion matrix for a predictive model?
R multilevel mediation with uneven sample sizes in Y and M models
I'm trying to run a multilevel mediation analysis in R. I get the error: Error in mediate(model.M, model.Y, treat = "treat", mediator = mediator, data=data): number of observations do not match between mediator and outcome models Models M and Y are multilevel lme4 models, and there are uneven sample sizes in these models. Is there anything I can do to run this analysis? Will it really only run if I have the same sample sizes in each model?
Fit the model with less observations first (which is, I guess, model.Y, because that model has more predictors and thus is more likely to have more missings), then use the model frame from that model as data for the 2nd model: model.M <- lmer(..., data = model.Y#frame) That should work.
R: Linear regression model does not work very well
I'm using R to fit a linear regression model and then I use this model to predict values but it does not predict very well boundary values. Do you know how to fix it? ZLFPS is: ZLFPS<-c(27.06,25.31,24.1,23.34,22.35,21.66,21.23,21.02,20.77,20.11,20.07,19.7,19.64,19.08,18.77,18.44,18.24,18.02,17.61,17.58,16.98,19.43,18.29,17.35,16.57,15.98,15.5,15.33,14.87,14.84,14.46,14.25,14.17,14.09,13.82,13.77,13.76,13.71,13.35,13.34,13.14,13.05,25.11,23.49,22.51,21.53,20.53,19.61,19.17,18.72,18.08,17.95,17.77,17.74,17.7,17.62,17.45,17.17,17.06,16.9,16.68,16.65,16.25,19.49,18.17,17.17,16.35,15.68,15.07,14.53,14.01,13.6,13.18,13.11,12.97,12.96,12.95,12.94,12.9,12.84,12.83,12.79,12.7,12.68,27.41,25.39,23.98,22.71,21.39,20.76,19.74,19.49,19.12,18.67,18.35,18.15,17.84,17.67,17.65,17.48,17.44,17.05,16.72,16.46,16.13,23.07,21.33,20.09,18.96,17.74,17.16,16.43,15.78,15.27,15.06,14.75,14.69,14.69,14.6,14.55,14.53,14.5,14.25,14.23,14.07,14.05,29.89,27.18,25.75,24.23,23.23,21.94,21.32,20.69,20.35,19.62,19.49,19.45,19,18.86,18.82,18.19,18.06,17.93,17.56,17.48,17.11,23.66,21.65,19.99,18.52,17.22,16.29,15.53,14.95,14.32,14.04,13.85,13.82,13.72,13.64,13.5,13.5,13.43,13.39,13.28,13.25,13.21,26.32,24.97,23.27,22.86,21.12,20.74,20.4,19.93,19.71,19.35,19.25,18.99,18.99,18.88,18.84,18.53,18.29,18.27,17.93,17.79,17.34,20.83,19.76,18.62,17.38,16.66,15.79,15.51,15.11,14.84,14.69,14.64,14.55,14.44,14.29,14.23,14.19,14.17,14.03,13.91,13.8,13.58,32.91,30.21,28.17,25.99,24.38,23.23,22.55,20.74,20.35,19.75,19.28,19.15,18.25,18.2,18.12,17.89,17.68,17.33,17.23,17.07,16.78,25.9,23.56,21.39,20.11,18.66,17.3,16.76,16.07,15.52,15.07,14.6,14.29,14.12,13.95,13.89,13.66,13.63,13.42,13.28,13.27,13.13,24.21,22.89,21.17,20.06,19.1,18.44,17.68,17.18,16.74,16.07,15.93,15.5,15.41,15.11,14.84,14.74,14.68,14.37,14.29,14.29,14.27,18.97,17.59,16.05,15.49,14.51,13.91,13.45,12.81,12.6,12,11.98,11.6,11.42,11.33,11.27,11.13,11.12,11.11,10.92,10.87,10.87,28.61,26.4,24.22,23.04,21.8,20.71,20.47,19.76,19.38,19.18,18.55,17.99,17.95,17.74,17.62,17.47,17.25,16.63,16.54,16.39,16.12,21.98,20.32,19.49,18.2,17.1,16.47,15.87,15.37,14.89,14.52,14.37,13.96,13.95,13.72,13.54,13.41,13.39,13.24,13.07,12.96,12.95,27.6,25.68,24.56,23.52,22.41,21.69,20.88,20.35,20.26,19.66,19.19,19.13,19.11,18.89,18.53,18.13,17.67,17.3,17.26,17.26,16.71,19.13,17.76,17.01,16.18,15.43,14.8,14.42,14,13.8,13.67,13.33,13.23,12.86,12.85,12.82,12.75,12.61,12.59,12.59,12.45,12.32) QPZL<-c(36,35,34,33,32,31,30,29,28,27,26,25,24,23,22,21,20,19,18,17,16,36,35,34,33,32,31,30,29,28,27,26,25,24,23,22,21,20,19,18,17,16,36,35,34,33,32,31,30,29,28,27,26,25,24,23,22,21,20,19,18,17,16,36,35,34,33,32,31,30,29,28,27,26,25,24,23,22,21,20,19,18,17,16,36,35,34,33,32,31,30,29,28,27,26,25,24,23,22,21,20,19,18,17,16,36,35,34,33,32,31,30,29,28,27,26,25,24,23,22,21,20,19,18,17,16,36,35,34,33,32,31,30,29,28,27,26,25,24,23,22,21,20,19,18,17,16,36,35,34,33,32,31,30,29,28,27,26,25,24,23,22,21,20,19,18,17,16,36,35,34,33,32,31,30,29,28,27,26,25,24,23,22,21,20,19,18,17,16,36,35,34,33,32,31,30,29,28,27,26,25,24,23,22,21,20,19,18,17,16,36,35,34,33,32,31,30,29,28,27,26,25,24,23,22,21,20,19,18,17,16,36,35,34,33,32,31,30,29,28,27,26,25,24,23,22,21,20,19,18,17,16,36,35,34,33,32,31,30,29,28,27,26,25,24,23,22,21,20,19,18,17,16,36,35,34,33,32,31,30,29,28,27,26,25,24,23,22,21,20,19,18,17,16,36,35,34,33,32,31,30,29,28,27,26,25,24,23,22,21,20,19,18,17,16,36,35,34,33,32,31,30,29,28,27,26,25,24,23,22,21,20,19,18,17,16,36,35,34,33,32,31,30,29,28,27,26,25,24,23,22,21,20,19,18,17,16,36,35,34,33,32,31,30,29,28,27,26,25,24,23,22,21,20,19,18,17,16) ZLDBFSAO<-c(1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2) My model is: fit32=lm(log(ZLFPS) ~ poly(QPZL,2,raw=T) + ZLDBFSAO) results3 <- coef(summary(fit32)) first3<-as.numeric(results3[1]) second3<-as.numeric(results3[2]) third3<-as.numeric(results3[3]) fourth3<-as.numeric(results3[4]) fifth3<-as.numeric(results3[5]) #inverse model used for prediction of FPS f1 <- function(x) {first3 +second3*x +third3*x^2 + fourth3*1} You can see my dataset here. This dataset contains the values that I have to predict. The FPS variation per QP is heterogenous. See dataset. I added a new column. The fitted dataset is a different one. To test the model just write exp(f1(selected_QP)) where selected QP varies from 16 to 36. See the given dataset for QP values and the FPS value that the model should predict. You can run the model online here. When I'm using QP values in the middle, let's say between 23 and 32 the model predicts the FPS value pretty well. Otherwise, the prediction has big error value.
Regarding the linear regression model I should use Weighted Least Squares as a Solution to Heteroskedasticity of the fitted dataset. For references, see here, here and here. fit32=lm(log(ZLFPS) ~ poly(QPZL,2,raw=T) + ZLDBFSAO, weights=1/(1+0.5*QPZL^2)) The other code remains the same. This model gives me lower prediction error than the previous.
Linear Discriminant Analysis in R - Training and validation samples
I am working with lda command to analyze a 2-column, 234 row dataset (x): column X1 contains the predictor variable (metric) and column X2 the independent variable (categorical, 4 categories). I would like to build a linear discriminant model by using 150 observations and then use the other 84 observations for validation. After a random partitioning of data i get x.build and x.validation with 150 and 84 observations, respectively. I run the following fit = lda(x.build$X2~x.build$X1, data=x.build, na.action="na.omit") Then I run predict command like this: pred = predict(fit, newdata=x.validation) From the reading of the commands description I thought that in pred$class I would get the classification of validation data according to the model built, but I get the classification of 150 observations instead of the 84 I intended to use as validation data. I don't really know what is happening, can someone please give me an example of how I should be conducting this analysis? Thank you very much in advance.
Try this instead: fit = lda(X2~X1, data=x.build, na.action="na.omit") pred = predict(fit, newdata=x.validation) If you use this formula x.build$X2~x.build$X1 when you build the model, predict expects x.build$X1 column in the validation data. Obviously there isn't one so you get prediction for training data.