Connect your moderator Slack workspace to receive post notifications:
Sign in with Slack

Few data, how to squeeze it to the fullest?

Hello all!

In our project each datapoint is the result of an experiment done in the lab (some of these take 90 days) so obviously we don't have thousands of points. After some reading and reserch we are aware of the importance when you have few data of being extrimely careful with overfitting, identifying outliers, giving priority to simple models, validate with CV leave one out...

The questions are, which techniques are the best for handling the lack of data? Can we really trust leave one out CV or is there a better way to estimate the error of our model?

The lab is also interested in us creating confidence intervals (or other measurements to give reliability for their experiments) for the data they have provided us. Could you give us some references (code libraries and theory) where we could start reading?

Thank you in advance :)

Page 1 of 1

Add comment

Post as Anonymous Dont send out notification