|
That seems like overfitting. Not knowing about the data here, I'd start by checking to make sure your data is homogeneous: that you don't have a huge set of similar data in training and then a big set of different ones in test. This can happen if you don't randomly separate the two categories and flat copy the data into test where filenames might be grouping classes or similar samples within a class. Your test data could just be the tail end of your class listings. You could also be feeding very high res data, in which case, you would either need a lot more data, or you could just scale them at the input if that's not already happening. Augmentation could improve your acc, but we're talking single digit % typically. I'm mostly familiar with TF, but I imagine PyTorch would have that built into the front end as well. But this disparity points to some more fundamental issue with the data imo. e: I just saw that you're trying to ID cars by mfg. with CNN. Assuming it's a bunch of random images off the net at all sorts of different angles and views, that's going to be hard to get accurate at 2000 samples. But I think you should be able to see better test acc. than that. Lower input res and making sure they're uniformly distributed should really boost that. former glory fucked around with this message at 15:56 on Jun 23, 2022 |
# ¿ Jun 23, 2022 15:46 |
|
|
# ¿ May 21, 2024 02:07 |