SquishyQuokka
SquishyQuokka

When do you stop training a Neural Network?

  1. Early cut off looking at training_loss
  2. Early cut off looking at validation_loss
  3. Set Higher Regularization?

Increasing data is not an option, have used SMOTE related technqiues for interpolating more data for the imbalanced classes already.

Just trying to understand what do other folks do.

Post image
20mo ago
Jobs
One interview, 1000+ job opportunities
Take a 10-min AI interview to qualify for numerous real jobs auto-matched to your profile 🔑
+322 new users this month
GoofyDonut
GoofyDonut

When do you stop training a Neural Network

When it gains sentience

GroovyBoba
GroovyBoba

Before you max out your credit card due to compute bill.

SquishyQuokka
SquishyQuokka
Gojek20mo

This is an old project. Was looking through Tensorboard log for a run I did in 2021, during the pandemic.

GroovyBoba
GroovyBoba

Ideally you should look at validation loss as that is the true unseen data and would be closest to new unseen real world data.

FloatingRaccoon
FloatingRaccoon
Adidas20mo

Try 500 epochs

SquishyQuokka
SquishyQuokka
Gojek20mo

Overfitting

Discover more
Curated from across
Data Scientists
by SillyWaffleProduct Management

Indic focused SLM/LLMs

Is anyone here who has built Indic-focused SLM/LLM models for text generation? If yes, please DM me. Want to understand the following things

  1. Want to understand the training/fine-tuning and running costs.
  2. I also want to understand ...