Why Is It Important To Shuffle Your Dataset Before Training An ML Model
…And what happens if you don't.
ML models may fail to converge for many reasons. Here's one of them which many folks often overlook.
If your data is ordered by labels, this could negatively impact the model's convergence and accuracy. This is a mistake that can typically go unnoticed.
In the above demonstration, I trained two neural nets on the same data. Both networks had the same initial weights, learning rate, and other settings.
However, in one of them, the data was ordered by labels, while in another, it was randomly shuffled.
As shown, the model receiving a label-ordered dataset fails to converge. However, shuffling the dataset allows the network to learn from a more representative data sample in each batch. This leads to better generalization and performance.
In general, it's a good practice to shuffle the dataset before training. This prevents the model from identifying any label-specific yet non-existing patterns.
In fact, it is also recommended to alter batch-specific data in every epoch.
P.S. This was a mistake that I once made a few years back, and it never occurred to me that shuffling is that important.
What are your thoughts on this? Let me know :)
👉 Read what others are saying about this post on LinkedIn: Post Link.
👉 If you love reading this newsletter, feel free to share it with friends!
Find the code for my tips here: GitHub.
I like to explore, experiment and write about data science concepts and tools. You can read my articles on Medium. Also, you can connect with me on LinkedIn.
Hey,
In this, I couldn't understand what shuffling batch epochs meant. I did undergo a formal ML course, so I am familiar with the basic terminology. But I don't understand this batch epoch. Could you help me with that?
Thank you.