Replies: 1 comment
-
Small datasets aren't the best to work with DL models. Reducing the complexity of the model (using hyperparameters) is recommended when dataset is too small. And any engineered feature which would capture the domain knowledge would also help the model learn better.. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
As I've observed, models are often trained on fairly large datasets, for instance, using the load_covertype_dataset function, where the dataset contains over 580k samples. Currently, I am working on a problem involving the prognosis of injury severity for patients with traumatic brain injury (TBI). In my case, the dataset is quite limited, with only 1,500 samples and 70 features per sample. However, given that this is real-world data from patients in a hospital, this is actually a substantial amount of data.
I’ve tested all the models on my dataset, but I've noticed that they are extremely sensitive when trained on small datasets. Even a slight adjustment can result in significant changes to the model's outcomes. For example, when I change parameters like the number of layers in the
CategoryEmbeddingModel
or the number of attention blocks in theFTTransformer
, the model almost completely fails to learn anything. The loss and accuracy values remain unchanged, showing no improvement or variation at all.Is there any way I can improve the stability of models when working with small datasets? I would really appreciate any suggestions. Thank you very much!!
Beta Was this translation helpful? Give feedback.
All reactions