Skip to main content

Questions tagged [hyperparameter-tuning]

Hyperparameter tuning (also called hyperparameter optimization) refers to the process of finding the optimal set of hyperparameters for a given machine learning algorithm.

Filter by
Sorted by
Tagged with
0 votes
0 answers
16 views

I'm using Optuna to optimize LightGBM hyperparameters, and I'm running into an issue with the variability of best_iteration across different random seeds. Current ...
invalid syntax's user avatar
3 votes
1 answer
101 views

I have a dataset which I split into training, testing, and out-of-time sets. Then I feed my training set into K Fold CV. I understand that K Fold Cross Validation is used as a method to select the &...
user24758287's user avatar
0 votes
1 answer
21 views

Say I have two different models with different hyperparameters and I want to compare the performance of both models on some dataset. One model is much simpler than the other and, therefore, if I were ...
Frederico Portela's user avatar
3 votes
1 answer
41 views

What I’m trying to figure out I'm working on a machine learning project and would love to hear your thoughts on two things: A. How to prioritize feature exploration B. Whether to fix hyperparameters (...
Ten's user avatar
  • 31
1 vote
0 answers
11 views

I have used Hyperband automatic tuning for an ANN model to predict price. After running the model with the automatic tuning, I am obtaining an R2 score of 1.00 that suggests overfitting, however, I am ...
leakie's user avatar
  • 11
4 votes
0 answers
55 views

I am working on hyperparameter optimization for YOLOv8 using a metaheuristic algorithm. Currently, I am testing the ...
NewPartizal's user avatar
4 votes
3 answers
107 views

I read two articles by the same guy where he uses the whole dataset for hyperparameter optimisation using with CV and then evaluates the model with the best hyperparameters using leave one out on the ...
Lisana Daniel's user avatar
2 votes
0 answers
94 views

I recently observed something unexpected: Although monotonic feature transformation does not affect the performance of decision tree-based models with default hyperparameters, it actually does affect ...
user4924539's user avatar
0 votes
0 answers
44 views

GPT3 has several hyper-parameters that define the network architecture. My question is: which of these hyper-parameters, when increased, provide the most performance benefit vs computational cost? ...
chausies's user avatar
  • 121
1 vote
0 answers
52 views

I have an algorithm that trains a binary predictive model for a specified number of features from the dataset (features are all of the same type, but not all important.) Thus, the number of features ...
Roger V.'s user avatar
  • 117
1 vote
1 answer
138 views

I am a student and am looking for your help. I have two datasets, including pre-treatment CT scan and post-treatment CT scan. I want to compare these datasets to determine which yields the best ...
waleed almutairi's user avatar
1 vote
1 answer
105 views

I am running machine learning models (all with sci-kit learn estimators, no neural networks) using a custom dataset with a number of features and binomial output. I first split the dataset into 0.6 (...
Alex's user avatar
  • 11
1 vote
0 answers
33 views

I'm writing a python script for Orange Data Mining to plot the gaussian processes in order to find the best hyperparameters for the 5-FoldCrossValidation Accuracy metric. The three models are SVC, ...
Mattma's user avatar
  • 11
0 votes
1 answer
168 views

I am tuning the hyperparameter of the decision tree for a data set of 550 samples. As I am comparatively new in hyperparameter tuning(I am learning and implementing), I am confused about what values ...
h_ihkam's user avatar
  • 33
1 vote
0 answers
55 views

I have a dataframe containing 1324 rows and 28 columns and I'm kinda lost on which approach to go for when training regression models. Currently I perform a data split and run GridSearchCV to pick the ...
Davi Magalhães's user avatar
1 vote
0 answers
109 views

I am new to ML and currently working on improving the accuracy of an MLPClassifier in scikit. My code looks like so ...
Leandro's user avatar
  • 25
1 vote
0 answers
56 views

I am trying to train LSTM model (containing four LSTM layers (500 units each) and three droupouts and a fully connected output layer to do regression) on timeseries data. To start with, I tried to ...
Mahesha999's user avatar
1 vote
0 answers
28 views

I am trying to train LSTM model on a timeseries data with 1.6 million records. I have taken window size of 200. Initially I tried to overfit the model (train data = test data) on tiny dataset (few ...
Mahesha999's user avatar
1 vote
1 answer
55 views

I want to train an XGBoost model, and here's how I believe the process should go: Step 1: Find the optimal hyperparameters using GridSearchCV. Step 2: Evaluate the selected parameters. My question is: ...
Math_D's user avatar
  • 121
0 votes
1 answer
29 views

In Scikit Learn RandomSearch or GridSearch , how to include non hyper parameters in the tuning process?! Non hyper parameters are parameters not related to the machine learning algorithms. For example ...
Emad Ezzeldin's user avatar
2 votes
1 answer
189 views

I would like to ask for help with the following. Given the following dataset, which I have split into train and test sets: ...
ProgrammerGnome's user avatar
1 vote
1 answer
73 views

I am testing the effects of different layer counts/depths, unit counts, and layer types for natural language processing. I made a Kaggle notebook where I manually create different layers and then ...
Joachim Rives's user avatar
2 votes
1 answer
1k views

Most Text Generation Models use beam search to select the optimal output candidate. How does one choose the optimal beam size? It would probably vary from task to task, dataset to dataset, and model ...
Tathagato Roy's user avatar
1 vote
2 answers
641 views

I have about 30 variables and trying to create a Random Forest model. All the variables are expected to be predictors of outcome. I want to find the best model based on a C-stat score with any number ...
user2704338's user avatar
1 vote
1 answer
208 views

Assuming a standard train/validation/test split, the common practice is (a) to train multiple models with different hyper-parameter configurations on the training set, (b) to evaluate these models ...
Enk9456's user avatar
  • 135
1 vote
1 answer
47 views

Jane trains three different classifiers: Logistic Regression, Decision Tree, and Support Vector Machines on the training set. Each classifier has one hyper-parameter (regularisation parameter, depth-...
Tom's user avatar
  • 11
0 votes
1 answer
79 views

I have an ML pipeline built with DVC that I use for experiment tracking. This allows running and tracking several experiments. Also, using hydra integration I can grid search hyper parameters. However,...
giulatona's user avatar
1 vote
1 answer
745 views

The way I read almost lots of ML advice on these datasets sounds like "You train a model that's randomly chosen hyperparameters first on the training set, then you ignore this bit of the work, ...
Socorro's user avatar
  • 121
-1 votes
1 answer
76 views

I was hoping to hypertune my decisiontree model , however I keep running into this error: TypeError: DecisionTreeClassifier() got an unexpected keyword argument 'criterion' here what I tried: ...
John Adams's user avatar
1 vote
1 answer
99 views

The scikit-learn classes HalvingGridSearchCV and HalvingRandomSearchCV implement a hyperparameter tuning method known as successive halving. It is an iterative selection process in which all the ...
Evan Aad's user avatar
  • 185
0 votes
1 answer
67 views

I am building a class that follows the workflow: Model Selection and Fitting The class accepts a list of models and their respective hyperparameter grids. It then performs a standard fitting process ...
Guilherme Raibolt's user avatar
2 votes
1 answer
109 views

I am currently learning binary classification. The problem is classifying positive and negative movie reviews. The dataset is 25,000 reviews with each review represented by 10,000 of the most used ...
Omer Mualem's user avatar
1 vote
2 answers
114 views

The description on the RandomizedSearchCV says this about best hyperparameters : "Estimator that was chosen by the search, i.e. estimator which gave highest ...
pppp_prs's user avatar
0 votes
1 answer
817 views

Is hyperparameter tuning done on training or validation data set? The post here gives mixed opinion as of whether the training set should be used for hyperparameter tuning. And I would like to know ...
Student's user avatar
  • 421
0 votes
1 answer
46 views

I'm towards the completion of my first data science project that will go into my GitHub portfolio. I'll be happy for some clarification regarding the machine learning models section: I got a little ...
Sigal Cohen's user avatar
0 votes
1 answer
819 views

When I use RandomizedSearchCV, if I put the random state I always obtain the same results with the same hyperparams trainer. So, is it mandatory to use? Because in my opinion it is better to always ...
Flavio Brienza's user avatar
0 votes
0 answers
131 views

If I use RandomizedSearchCV to find the optimal hyperparams of a model, can I create another model, with those parameters, to calibrate probabilities using CalibratedClassifierCV? The new model is not ...
Flavio Brienza's user avatar
2 votes
1 answer
889 views

I'm currently trying to build a model using CatBoost. For my parameter tuning, I'm using optuna and cross-validation and pruning the trial checking on the intermediate cross-validation scores. Here ...
GiusWestsideDS's user avatar
1 vote
1 answer
306 views

I want to ask you a question. Suppose I use the following RandomizedSearchCV to find the model's best hyperparams: ...
Flavio Brienza's user avatar
0 votes
1 answer
96 views

I am trying an example which I am training on a huge dataset 5M (only 4 features) rows with Cudf and CUml and I am using SGD logistic regression because I must predict if the patient if is sick or not ...
gkasap's user avatar
  • 1
1 vote
1 answer
145 views

Goal: Compare preprocessing methods, models, and hyperparameters without leaking into the final generalization estimate, applying cross-validation (cv), i.e. NOT applying any fixed train/test splits. ...
le8rning's user avatar
0 votes
1 answer
698 views

I am working on a classification problem using Sci Kit Learn and am confused on how to properly tune hyper parameters to get the "best" model. Before any tuning, my logistic regression ...
d0dg3r_k1d's user avatar
0 votes
1 answer
60 views

When looking to train a model, does it make sense to have a 60-20-20 train val test split, first hyper parameter tuning over the training dataset, using the validation set, picking the best model. ...
Socorro's user avatar
  • 121
1 vote
1 answer
886 views

I am using MLPClassifier from sklearn and I would like to tune it with GridSearchCV. But I don't know which set of values to include for hidden_layer_sizes, max_iter, activation, solver, etc. How can ...
Penjan's user avatar
  • 11
0 votes
0 answers
2k views

I'm getting this error when I try to use Keras Tuner with my model: Number of consecutive failures excceeded the limit of 3. .... KeyError: 'mean_squared_error' Here's my code: ...
soilwatch's user avatar
1 vote
1 answer
29 views

My question is simple and yet quite hard to find an answer to. In an unsupervised method, for example, when you have to reconstruct an input, how can you tell if your loss is good enough? Generally, ...
BilboBuggins's user avatar
0 votes
2 answers
1k views

The XGBoost library has its own implementation of cross validation through xgboost.cv(). It looks like it requires data be stored as a DMatrix. Instead of using <...
Eli's user avatar
  • 101
0 votes
2 answers
72 views

I want to perform a global optimization of the entire model development pipeline. I have several stages of development, each of which can be performed automatically: preprocessing, removal of outliers/...
Andrew's user avatar
  • 406
2 votes
2 answers
3k views

I'm testing hyperparameters for an SVM, however, when I resort to Gridsearch or RandomizedSearchCV, I haven't been able to get a resolution, because the processing time is exceeding hours. My dataset ...
Paulo Sergio Moreira's user avatar

1
2 3 4 5 6