Discussion about this post

User's avatar
Neural Foundry's avatar

The age example nails why discretization works, treating 29 vs 30 the same as 59 vs 60 misses the behavioral inflection points entirely. One thing worth adding is that equal-frequency binning can backfire with skewed distributions where you end up grouping meaningfully diferent values just to maintain bin population. I've seen this with income data where the top bin becomes useless. The overfitting risk you mentioned is real but I think it's less about dimensionality and more about leakage, discretizing on training data then applying those bins to test can create artifical signal.

No posts

Ready for more?