Questions tagged [accuracy]
Accuracy of an estimator is the degree of closeness of the estimates to the true value. For a classifier, accuracy is the proportion of correct classifications. (This second usage is not good practice. See the tag wiki for a link to further information.)
846 questions
3
votes
1
answer
118
views
Accuracy in Machine Learning vs. Accuracy in Statistics vs. pass@1,1 in Generative Modeling: What's the Difference?
I've encountered the term "accuracy" used differently across several evaluation contexts, and I want to clearly understand their mathematical and conceptual distinctions using consistent ...
0
votes
2
answers
52
views
How to investigate if my poor classification is because of bad data or some other reason [duplicate]
I currently have a RandomForestClassifier that is classifying workload based on fNIRS data. Our classification accuracy is about 49% I want to investigate why our classification accuracy is so bad and ...
1
vote
1
answer
43
views
Two approaches to go from 2AFC accuracy to d′ - how do they differ and which should I use?
I’ve recently encountered two approaches used to express performance on perceptual tasks as d' when trying to convert (non-linear) accuracy on a 2AFC (2-alternative forced choice) task to a linear ...
7
votes
1
answer
197
views
Doubling your accuracy - extension
Frederick Mosteller's 50 Challenging Problems in Probability has a nice question I have not seen before, and I was wondering whether it could be extended.
49. Doubling your accuracy
An unbiased ...
1
vote
0
answers
69
views
Order sensitivity of scoring rules
This is from another question here.
The theorem below is from Lambert's paper about forecasting, (Elicitation and Evaluation of Statistical Forecasts):
$\textbf{Proposition}\quad 1:$ Let $(\Theta = \{\...
3
votes
1
answer
126
views
Calculation of geometric mean for classification
Consider binary classification, the geometric mean is defined as $\sqrt{\text{Precision} \times \text{Recall}} = \sqrt{ \frac{TP}{TP+FP} \times \frac{TP}{TP+FN} }$. But there can be different TP/FP/FN ...
1
vote
0
answers
75
views
How to measure accuracy between multiple raters and a reference value?
I am interested in assessing the accuracy of raters to a reference standard for subjective ratings on a Likert scale from 1-10 as in:
...
3
votes
1
answer
228
views
How do I calculate Harrell's c statistic for a Royston Parmar model?
I am trying to calculate the concordance (c) statistic for a Royston-Parmar model. My model stratifies the baseline hazard and uses splines to model log(t).
I am not sure If I am calculating the c-...
0
votes
0
answers
80
views
Is using the TEST set to calculate the optimal threshold for binary classification and then calculating the accuracy on the same test set wrong
I have a dataset that has been split into 2 parts, train and test set. After training a model with the training set to classify between class 0 and 1, I used the sklearn roc_curve to calculate the ...
2
votes
1
answer
86
views
Metric choice for Machine Learning algorithm
I am currently building a ML model for a binary classification problem.
I am currently using a curated dataset that was provided in a research paper, that has been perfectly balanced. However, it is ...
0
votes
0
answers
33
views
Evaluating Accuracy of mixture model clustering and categorisation
I am running a Mixture model and I have no free parameters, I just have it evaluating for a given datapoint, its likelihood of belonging to one cluster. Separately, I have a ground truth about these ...
2
votes
3
answers
153
views
Testing forecasting accuracy - outliers [ with example]
I have a simple model that produces forecast values. The model works on hourly data. Now, I am only interested in observations with flags. I would like to identify where the forecasts are ...
1
vote
2
answers
149
views
Is it possible that false-positive rate decreases with increasing prevalence?
I am interested in the effect of prevalence on prediction performance. Chouldechova (2016) states that:
[w]hen using a test-fair [recidivism prediction instrument] in
populations where recidivism ...
1
vote
1
answer
134
views
How to evaluate performance of classification model for different subsets of classes?
Consider a classification problem where there are N classes. While this may seem strange, I have a model that processes features, and essentially, evaluate which classes are impossible (or near ...
0
votes
0
answers
34
views
assessing classifier accuracy when class presence is scarce
What can I do, to assess a classifiers accuracy, when class presence is scarce.
Setup 1: I have 1000 boxes, 500 contain gold. I build an automated tool to find the gold.
The recommended approach would ...
0
votes
1
answer
105
views
Should I use ROC curve in my training set after training a Random Forest classification model with k-fold cross validation?
I have a conceptual question: after dividing a dataset into a training and test set (70:30), both are balanced and shuffled, should I use the Confusion Matrix and the ROC curve of a model generated by ...
1
vote
1
answer
63
views
The accuracy of the cut-off is significantly dependent on an independent continuous variable
Say one continuous variable differentiates between disease and nondisease quite accurately, but as people progress in age, this variable becomes less accurate. Is there a way to determine the accuracy ...
7
votes
3
answers
1k
views
How can I explain the difference in accuracies in different ML models?
I have applied various ML models (fundamental and ensemble) to the same dataset for classification problem solving.
AdaBoost, Bagging, and XGBoost classifiers gave the best accuracies. However, they ...
0
votes
0
answers
69
views
Comparing performance of probabilistic regression models - how to adapt Brier score?
Suppose I have two predictions models, Model 1 and Model 2. I have a dataset containing observations, features and actual outcomes. For each observation, the “outcomes” (i.e. predictions) that the ...
0
votes
0
answers
108
views
How to measure Statistical Significance for calculated Sensitivity, Specificity, Precision, Accuracy and f1 values?
I am trying to understand how to calculate one or more measures of statistical significance to display alongside metrics I've calculated from my data.
Abbreviations I am using in the rest of this post:...
1
vote
1
answer
155
views
Metrics that weighs under forecasting and over forecasting differently
I have multiple moving averages forecasts that use different look back periods. I’m measuring accuracy using MAPE. Out of all the options, I want to select the best performing moving average. However, ...
11
votes
1
answer
3k
views
Getting 99-100% accuracy on my training/validation data but performs bad on completely new data
I have a large dataset of the ASL (American Sign Language). I split this data into 70:15:15 for train, validation, test.
I then trained a CNN model on it, where I trained using the 70%, and evaluated ...
3
votes
1
answer
109
views
Which is the denominator of the Brier score for joint multiple variables predictions?
Brier score can be computed for joint predictions of multiple variables, each with multiple categories.
Let's say we have 4 variables with 3 possible classes each.
In that case, the denominator of the ...
1
vote
1
answer
84
views
Accuracy "overfits" but loss doesn't?
I'm perplexed as to why my loss doesn't go up when the accuracy goes down (after about 40 epochs). Isn't it possible to tell overfitting from the loss curve alone? (I'm of course referring the ...
1
vote
0
answers
106
views
Help interpreting multi-class confusion matrix
I'm looking at the SAMHSA Mental Health Client-Level dataset. I did some t-SNE plots (dropping irrelevant cols, normalizing some, one-hot encoding some) of 500k rows out of 6.5mil.
I'm trying to do ...
3
votes
2
answers
630
views
Is F-score the same as accuracy when there are only two classes of equal size?
The title says it all: Is F-score the same as accuracy when there are only two classes of equal sizes?
For my specific case, I have measurements of a group of people under two different situations and ...
1
vote
1
answer
86
views
AdaBoost in Matlab - Only get 88% accuracy with Fisher's Iris data set while 100% in one layer Neural Network
I have made AdaBoost in Matlab. I get 88% in accuracy when I use Fisher's Iris flower set data.
Here is the working example:
...
1
vote
1
answer
114
views
Estimate sensitivity and specificity from inter-rater agreement?
Assuming two binary (Y in {0, 1}) annotators or classifiers (A and B), that are:
Conditionally independent, i.e. P(A=0, B=0|Y=1) = P(A=0|Y=1)*P(B=0|Y=1) and the same for Y=0.
Better than random, i.e. ...
1
vote
1
answer
1k
views
Symmetric AND Weight MAPE Calculation
I'm responsible to forecast a portfolio of consumer products on a monthly basis, and in calculating forecast accuracy, I'm lead to the MAPE (Mean Average Percent Error), which is useful, but has, ...
1
vote
1
answer
127
views
Cartesian precisions from spherical standard deviations
I have to simulate a simple sensor, which has 3 standard deviations defined in spherical frame : sigma-azimuth, sigma-elevation, sigma-distance.
When I simulate a detection, I compute a noisy position ...
0
votes
0
answers
89
views
Independent features but PCA improves classifiers accuracy significantly. Why?
that's my first question on here :)
I am working with the kNN classifier on datasets from the multivariate normal distribution. I have to groups coming from ...
0
votes
0
answers
112
views
Good training/test results, but very poor performance in inference as the stream data is coming
I am interested in the audio classification problem. After labeling the audio recordings I have in Praat software environment, I extract the MFCC features from each labeled frame and create an SVM ...
1
vote
1
answer
159
views
Finding a source for the definition of "clustering accuracy"
In papers about unsupervised clustering I see a lot of references to a metric "clustering accuracy" or "unsupervised clustering accuracy" (ACC) which is usually defined as ...
2
votes
1
answer
490
views
Accuracy score change a lot by changing random seed in train/test split
I'm running a ML algorithm on some data, and I noticed that if I change the random state inside the train_test_split function, accuracy score change in a quite wide range.
For example, with random ...
0
votes
0
answers
55
views
Why my XGBoostClassifier model results in perfect accuracy despite dropping corelated features?
I am trying to do a binary classification on ticket canceling data from kaggle.
I know this question has been asked before. For example here and here
Summary of what I learned in those references:
...
3
votes
2
answers
132
views
What is the most responsible way to visualise non-response / NA data in a plot?
I'm working on plotting census data, which has a fairly high non-response rate for some questions (5% or higher). This could actually shift the way we interpret the results in quite significant ways (...
2
votes
1
answer
586
views
Poor balanced accuracy and minority recall but perfect calibration of probabilities? Imbalanced dataset
I have a dataset with a class imbalance in favour of the positive class (85% occurence)
I'm getting a fantastically calibrated probabilities profile but balanced accuracy is 0.65 and minority recall ...
1
vote
1
answer
797
views
Distribution of accuracy from randomly guessing
Let's consider a true classification problem, that is, one where the predictor makes categorical predictions (not probabilities).
It makes sense to assess the accuracy of such a predictor. However, ...
6
votes
3
answers
3k
views
Is it okay to say that 95% confidence interval is more significant than 80%?
So, the higher the confidence interval the lower the false positive rate, but the false negative rate will increase lowering the recall.
Is it possible to determine which confidence interval is better/...
2
votes
0
answers
169
views
When *is* classification accuracy the right measure of performance
Plenty has been discussed on Cross Validated about the drawbacks of classification accuracy when it comes to evaluating classification models. One good answer is here, for instance.
Under what ...
1
vote
0
answers
13
views
Improve accuracy randomForest classification model [duplicate]
How do I improve the accuracy of the following data. It is from the following Kaggle competition which I am doing (despite it being closed for a school project).
...
1
vote
0
answers
52
views
PLSR: trait vs spectroscopic data gives very low R2 on plsr model in R
here is the sample data.
I have spectroscopy data as X-variables (from X1 to X80) and corresponding Y variable.
I need to run plsr model in R using "pls" package. There are two sheets. In ...
0
votes
0
answers
87
views
Did I lose precision when I applied inverse normalization and inverse standarization when I predict the new data in a regression problem?
Let's suppose that I'm trying to predict a stochastic forecast with machine learning models, and I don't have missing, null/NaN values and outliers. Also suppose that there is an error for the ...
1
vote
2
answers
148
views
why does a model with a larger val loss produce higher accuracy than a model with a smaller val loss?
I did ANN classification on training data with oversampling and without oversampling. For each data, the smallest validation loss is sought with trial and error of 18 models. In the data without ...
0
votes
1
answer
119
views
Looking for a formula to determine accuracy of a sample
I am looking for a formula to help me determine the accuracy of a population.
Here is my business problem.
I have about 1 million scanned documents of many types that are currently unclassified ...
0
votes
0
answers
64
views
p-value for results of sub group analysis with ML?
I developed a ML algorithm (Xgboost) to predict a target in my data set.
I obtain here the results of my predictions on my test set :
...
0
votes
0
answers
72
views
Accuracy on NN model decrease after random oversampling using library ROSE
I did random oversampling to handle unbalanced positive and negative data. When I didn't do random oversampling, the accuracy I got was 88%, when I oversampled the train data, it got 87% accuracy and ...
1
vote
1
answer
148
views
How does someone achieve a desired confidence / accuracy when measuring using uncalibrated instrument?
I have an instrument that measures a value.
It is only possible to measure the value once i.e. the experiment can't be repeated (think recording a car's speed as it drives past).
The instrument is not ...
1
vote
0
answers
226
views
Confidence intervals for binary classification
I'm doing binary classification in Python with an SVM classifier, and I implemented stratified repeated cross validation to have more robust results.
I would like to calculate confidence intervals for ...
2
votes
1
answer
142
views
Upper bound on classification performance
Given a set of 128x128 images from three classes, I obtained an accuracy of 50% with a SVM on the flattened images (16384 'features').
Is this an upper bound on the performance of a SVM using any ...