Skip to main content

Questions tagged [ensemble-learning]

In machine learning, ensemble methods combine multiple algorithms to make a prediction. Bagging, boosting and stacking are some examples.

Filter by
Sorted by
Tagged with
1 vote
1 answer
94 views

I am looking to apply a calibration/correction approach on a set of sensors and I just wanted to know that the approach I am going to use is statistically correct and acceptable. I am using a set of ...
Milad's user avatar
  • 157
1 vote
0 answers
60 views

I am performing binary prediction on a dataset which contains missingness, and so I am leveraging Multiple Imputation (MI). For example, creating a train-test split, I perform MI on the training data ...
benedictjones's user avatar
0 votes
0 answers
46 views

Context I'm trying to create an Ensemble survival neural network with a custom loss function which consist of 3 base models, Random Survival Forest (RSF), Gradient Boosting Survival Model (GBSM) and a ...
Yugan Gogul Muthukumar's user avatar
1 vote
1 answer
66 views

I have this doubt, in the case an ensemble of linear basic models I am convinced (but I do not know the exact mathematical explanation for this) that the resulting ensamble works only for a linear ...
Gianni's user avatar
  • 173
1 vote
0 answers
74 views

I am working on an experiment with a dataset, where I compared the performance of stacking, blending, and voting using base models and logistic regression as my meta model. Although, voting seems to ...
0 votes
0 answers
129 views

I have three classes {-1, 0, 1}. The data is in the ratio 1:20:1 on average for the corresponding classes. I want to achieve High precision(>70%) and average recall (30%-40%) on classes -1 and 1. ...
CuriousRabbit's user avatar
1 vote
1 answer
116 views

I have two machine learning models for predicting some continuous variable $y$, say $y=f_1(X_1, \theta_1)$ and $y=f_2(X_2, \theta_2)$, and these models are of the same type (ANN). $X_1$ and $X_2$ ...
tunar's user avatar
  • 563
2 votes
1 answer
207 views

I have several (n) different forecasts of comparable quality for a variable, based on the same data but using wildly different statistical models. For each, I have generated an estimate for m periods ...
andrewH's user avatar
  • 3,297
4 votes
1 answer
902 views

I'll try to be brief. I have two questions about what exactly happens when I train a gradient boosted ensemble of trees using, say, XGBoost in order to perform a Gamma regression. I apologize in ...
user412834's user avatar
0 votes
1 answer
164 views

For a regression problem, I want to train an ensemble of deep neural networks to predict the labeled output as well as the uncertainty, similar to the approach presented in the paper Simple and ...
qubit's user avatar
  • 1
1 vote
0 answers
251 views

Given a random forest (or any other ensemble) where each of the $i=1..n$ trees/base estimators is trained by minimizing the mean squared error, then each tree/base estimator prediction $\hat{Y}_i(x) =...
Ggjj11's user avatar
  • 2,345
1 vote
0 answers
73 views

I am building a model to predict a customer purchase event on a website. Specifically for those customers who, overnight when the model is run, have not yet purchased. Prediction is important, but ...
Jon's user avatar
  • 141
1 vote
0 answers
74 views

When using an XGB model in the context of binary classification, I observed that the test estimates given by predict_proba were close but not equal to the results I ...
Juan Felipe Salamanca Lozano's user avatar
1 vote
3 answers
200 views

I have a dataset that contains a few patterns of missingness. For this dataset, I have a training set that is complete and contains all input features. My test set has complete observations for the ...
Squan Schmaan's user avatar
3 votes
1 answer
388 views

I am currently trying to build a stacked algorithm in order to determine how many people in each region of a country will be likely to buy a product versus its competitors. I have some data from an ...
huntercallum's user avatar
0 votes
0 answers
68 views

I am running an ensemble random forest model (a newer method published in 2020). The model works by using a double bootstrapping step to balance imbalanced training data. Then you grow multiple ...
Greatwhite4's user avatar
0 votes
1 answer
144 views

You are working on a binary classification problem with 3 input features and have chosen to apply a bagging algorithm (Algorithm X) on this data. You have set max_features = 2 and n_estimators = 3. ...
Tanjim Taharat Aurpa's user avatar
1 vote
0 answers
90 views

Let's say we want to stack 2 base models: an XGBoost regressor and a deep neural network by linearly combining their predictions as ...
statnoob's user avatar
1 vote
0 answers
51 views

I am trying to do some forecasts. I have produced multiple forecasts by a variety of methods. All of the forecasts I have generated so far have residuals that are strongly positively correlated.I ...
andrewH's user avatar
  • 3,297
0 votes
0 answers
81 views

When I train deep models with cosine annealing learning-rate scheduling and warm restarts, I get models that achieve completely different scores on my validation set, after each training cycle. There ...
SomethingSomething's user avatar
1 vote
0 answers
55 views

Many well-known methods address specific forms of model uncertainty that can be adjudicated empirically. For example, if we are fitting a predictive model and there is uncertainty about the set of ...
half-pass's user avatar
  • 3,850
4 votes
0 answers
150 views

I understand that random forests yield better performance than standard decision trees, but are less interpretable, because they do not generate a single tree. In this question, several users provided ...
nka5we's user avatar
  • 49
1 vote
0 answers
149 views

Assume I have a fancy procedure $w: X \to \mathbb{R}$ to come up with weights for examples $x \in X$. Think of it as similar to the weights used in e.g. some boosting procedures. Now, I want to build ...
ngmir's user avatar
  • 339
2 votes
1 answer
104 views

I'd like to check if there are correlation among predicted probabilities of models in a voting classifier. According to the table below, one of models, Model5, has mean 40.9% and standard deviation 46....
KAI's user avatar
  • 23
3 votes
1 answer
1k views

I've been looking into ways to estimate uncertainty for regression tasks on neural networks. One of the obvious options is ensemble modeling. Consider an ensemble of neural networks that all have ...
noNameTed's user avatar
  • 287
2 votes
1 answer
106 views

Lets assume I have 24 random forest models. Each of 24 random forest models produces a class prediction. I am currently using simple majority voting to select final prediction. Can someone please ...
MSKO's user avatar
  • 71
6 votes
1 answer
847 views

Can someone tell me what the name is for the following process? I have some data with inputs $x_i$ and outputs $y_i$, and I fit a simple model (e.g. linear regression) to them. Then, I compute the ...
Fai Wang's user avatar
1 vote
1 answer
677 views

I'm currently immersed in a challenging forecasting project centred around predicting the required work hours to complete various tasks within a team setting. My dataset comprises crucial attributes, ...
Tirth's user avatar
  • 113
1 vote
0 answers
38 views

According to Elements of Statistical Learning Ch 8.8, we can apply least sqaures at the population level to show that for a regression ensemble $f_1(x), f_2(x), \ldots , f_M(x)$ where $f_j: \mathbb{R}^...
Seraf Fej's user avatar
  • 556
2 votes
0 answers
209 views

I have two different CNNs trained on the same dataset. One performs a bit better than the other but I believe each can provide different and useful information. I use ...
raquelhortab's user avatar
1 vote
1 answer
107 views

Consider an ensemble of weak learners (i.e. regressors or classifiers) whose predictions are aggregated (e.g. via averaging or majority vote) into an ensemble estimate. This gives rise to the question ...
ngmir's user avatar
  • 339
2 votes
0 answers
104 views

In 'big data' settings where the number of samples $n$ may be very large (for fixed number of features), is bagging less or more effective at reducing variance? I heard the claim that it is less ...
WeakLearner's user avatar
  • 1,631
4 votes
1 answer
2k views

I predict a continuous variable by taking the average of $N$ model predictions. The models are different in terms of their functional form, i.e. a tree model, a neural net, etc. Is the average SHAP ...
shenflow's user avatar
  • 1,149
1 vote
0 answers
40 views

Background I am trying to create a model that can predict Type 2 diabetes in a patient based on MRI scans of their thigh muscle. Previous literature has shown that fat deposition in the muscle of ...
Saminy Creed's user avatar
1 vote
1 answer
165 views

I must identify if a data point is an outlier or not in a dataset (we don't have labels). I have different unsupervised models to identify the outlier. Then, I normalize the outlier score and I ...
AdrianJoi's user avatar
0 votes
0 answers
98 views

I am using catboost regressor and lgbm regressor to perform regression on dataset. I want to know the assumptions of both the models. Where can I find assumptions for both the models? Next I want to ...
Lopez's user avatar
  • 147
0 votes
2 answers
644 views

I'm working on a project classifying patients as having (1) or not having (0) a particular condition. Someone I work with has suggested fitting a decision tree on this data, and using the leaf node ...
k13's user avatar
  • 57
2 votes
2 answers
731 views

So it sounds like boosting techniques (eg. XGBoost) uses weak learners (stumps) to gradually learn sequentially. This is not in dispute I hope. However, with bagging techniques (eg. Random Forest) I'm ...
Katsu's user avatar
  • 1,051
2 votes
1 answer
158 views

Extremely confused about the following: Lets say we start out with a dumb weak learner. Since its the 0th model and hasnt learned anything yet, we have a high residual, lets say of 10,000. We produce ...
Katsu's user avatar
  • 1,051
1 vote
1 answer
665 views

Per my understanding, there are 2 kinds of "subsets" that can be used when creating trees: 1) Subset of the dataset, 2) Subset of the features used per split. The concepts that I'm comparing ...
Katsu's user avatar
  • 1,051
1 vote
1 answer
518 views

I have been reading up on super learner ensemble methods that utilize multiple models and model configurations to make model predictions as good or better than any individual base model previously ...
Jake Niederer's user avatar
0 votes
0 answers
55 views

As the title says, I was wondering, if I have to train every neural network of a deep ensemble on a different training set or on the same one. I ask this question because I am getting weird results. ...
Alucard's user avatar
  • 325
0 votes
0 answers
68 views

I'm struggling to find some clear math behind ensemble learning. I can simulate it very easily, eg: ...
Blaze's user avatar
  • 73
5 votes
1 answer
223 views

Regarding boosting in the context of machine learning. One definition I have encountered talks about turning multiple weak learners into one strong learner, and another talks about starting with a ...
A. Maman's user avatar
  • 222
1 vote
0 answers
42 views

I'm currently working on an ensemble of 5 differently trained networks using MinkLoc3D v2 as base-net. I'm currently investigating the reason for lousy recall when I compare the extracted database ...
JackFrost67's user avatar
0 votes
0 answers
190 views

What happens to the accuracy of a decision tree when it is pruned? Can be higher than the accuracy of the fully-grown decision tree?
san's user avatar
  • 11
0 votes
0 answers
132 views

Consider model A, a deployed model that produces a probabilty of if an event occur or not for a population. This I want to improve by building another model, model B, on top of model A. Model B should ...
Henri's user avatar
  • 253
3 votes
1 answer
579 views

Under Binary classification situation, error between function $f$ and basic learner(classifier) $h_i(x)$ is $$P(h_i(x)≠f(x))=\mathcal{E}.$$ It is assumed that $T$ basic classifiers are combined by a ...
Kombatant82's user avatar
3 votes
0 answers
52 views

Suppose that I wanted to choose from, for example, $Y = aX + \epsilon$ and $Y = aX^2 + \epsilon$. Is this meaningfully different from fitting $Y = a_1X + a_2X^2 + \epsilon$ and heavily penalizing $...
llllvvuu's user avatar
  • 223
0 votes
1 answer
71 views

I have a doubt from a passage in the ensemble learning chapter of Aurelien Geron's book "hands on machine learning... ". I do not understand If you do the math, you will find that the ...
Jose_Peeterson's user avatar

1
2 3 4 5
10