Nhabls

Nhabls t1_j32q6zp wrote

The "monopoly" is from the ecosystem mostly, not the hardware itself. Practicioners and researchers have a much better time using consumer/entry level professional nvidia hardware. So they use nvidia.

Mind you that in the supercomputer level there is no real "monopoly" as those people just develop their solutions from the ground up.

55

Nhabls t1_iwcdek4 wrote

The data doesn't seem that imbalanced, not to cause the issues you're having. And idk what you are using for augmentation but you can def augment classes to specifically solve imbalance ( i don't like doing that personally). My next guess would be looking at how you're splitting the data for train/val and/or freezing the vast majority of the pretrained model and maybe even just training on the last layer or 2 that you add on top.

Regardless, it's something that's useful to know (very frequent in real world datasets) here's a link that goes over how to weigh classes for such cases it's with tensorflow in mind but it's the same concept regardless

3

Nhabls t1_iwc3gcr wrote

What is the representation of each class? A class imbalance could create this exact behavior. You dont even need to use a data augmentation technique ( i don't have a particularly great opinion of them, personally) and just scale the weights appropriately instead.

Also what does "Standard" mean here?

2