Recent comments in /f/deeplearning
Tekno-12345 OP t1_jc6jl0q wrote
Reply to comment by HarissaForte in Using GANs to generate defective data by Tekno-12345
I have a lot of good images with no defects.
And very few images with defects.
In order to train a good object detector for this task, a huge amount of defects will be needed.
rushan3103 t1_jc6h2ih wrote
Reply to comment by Tekno-12345 in Using GANs to generate defective data by Tekno-12345
Have you looked into the paper MetaGAN ?
HarissaForte t1_jc6gder wrote
Reply to comment by Tekno-12345 in Using GANs to generate defective data by Tekno-12345
I don't get it… what do you have so far?
How many of
- image alone,
- image + labels,
- image + bbox,
- image + mask ?
Tekno-12345 OP t1_jc6drqj wrote
Reply to comment by HarissaForte in Using GANs to generate defective data by Tekno-12345
It could be..
However, such models require the data I am seeking
HarissaForte t1_jc66t3m wrote
Reply to Using GANs to generate defective data by Tekno-12345
> a model to detect defects
… as in object detection?
PM_ME_ENFP_MEMES t1_jc5yjdv wrote
Reply to comment by rezayazdanfar in How To Scale Transformers’ Memory up to 262K Tokens With a Minor Change? by rezayazdanfar
Own it bro, don’t listen to the haters. If your work is fabulous, call it fabulous, literally nobody can stop you and nobody that appreciates your work will care what you call it as long as it contributes to moving things forward.
qphyml t1_jc5y6ij wrote
Reply to comment by errgaming in Which topic in deep learning do you think will become relevant or popular in the future? by gokulPRO
Yes! Transforming regular data structures into graphs for analysis is a really powerful thing by itself. And then to utilize the powers in GNNs on top of that!
Tekno-12345 OP t1_jc5wf2t wrote
Reply to comment by Kuchenkiller in Using GANs to generate defective data by Tekno-12345
Thanks for the reply.
Yes classical methods did not generalize well on foreign data.
I don't have experience on GANs and wanted some experienced opinions.
I have some leads now, I'll try them out and get back to your comment if they did not work out.
Tekno-12345 OP t1_jc5vjac wrote
Reply to comment by gradientic in Using GANs to generate defective data by Tekno-12345
My problem with AD models is that they are very sensitive, even if trained on noise.
Any small diversion will be detected as defective.
Tekno-12345 OP t1_jc5vfvo wrote
Reply to comment by mcottondesign in Using GANs to generate defective data by Tekno-12345
Already tried these methods, they did not help much
rezayazdanfar OP t1_jc5ncpy wrote
Reply to comment by WallyMetropolis in How To Scale Transformers’ Memory up to 262K Tokens With a Minor Change? by rezayazdanfar
True but i didn't call my own work fabulous, i meant the main work. :)
WallyMetropolis t1_jc564sn wrote
Calling your own work 'fabulous' is a little unusual.
ForeskinStealer420 t1_jc4v00p wrote
Reply to comment by errgaming in Which topic in deep learning do you think will become relevant or popular in the future? by gokulPRO
This
Kuchenkiller t1_jc4tir0 wrote
Reply to comment by Kuchenkiller in Using GANs to generate defective data by Tekno-12345
Also, what is the reason for deep learning if you don't have the data? Have classical methods been unsuccessful?
Kuchenkiller t1_jc4t81a wrote
Reply to Using GANs to generate defective data by Tekno-12345
If you have enough images of defects but are just lacking the labeling (probably easier to come by) one approach is to generate random morphological structures on your bottles (e.g. just some random circles and ellipses) and then apply cycleGAN or CUT to transform from this "segmented" image domain to the domain of real images. As said, you still need a lot of data but don't need labelling. Just generating useful data from noise (basic GAN idea) can work in theory but is extremely hard to train. I had way more success with the domain transfer approach (my case in medical imaging)
ats678 t1_jc4lem7 wrote
Reply to Which topic in deep learning do you think will become relevant or popular in the future? by gokulPRO
In the same fashion as LLM, I think Large Vision Models and multimodal intersections with LLM are the next big thing.
Apart from that, I think things such as model quantisation and model distillation are going to become extremely relevant in the short term. If the trend of making models larger will keep running at this pace, it will be necessary to find solutions to run them without using a ridiculous amount of resources. In particular I can see people pre-train large multimodal models and then distill them for specific tasks
errgaming t1_jc4an9h wrote
Reply to Which topic in deep learning do you think will become relevant or popular in the future? by gokulPRO
Graph Neural Networks. My primary research area is currently in GNNs, and I believe it is very very underrated.
Dark-Penguin t1_jc49hd1 wrote
Reply to Which topic in deep learning do you think will become relevant or popular in the future? by gokulPRO
Evolutionary techniques
gradientic t1_jc430rz wrote
Reply to Using GANs to generate defective data by Tekno-12345
Not really direct answer to your question, but the general problem you are trying to solve is called image anomaly detection, there are well know approaches that try to solve this problem, some of them in an unsupervised manner (assuming that you have a significant dataset of images without anomalies - learn inlier look for outliers) - check https://towardsdatascience.com/an-effective-approach-for-image-anomaly-detection-7b1d08a9935b for some ideas and pointers (sry if pointing obvious things)
RoboiosMut t1_jc3ptw7 wrote
Reply to comment by GufyTheLire in Recommendations sources for Understanding Advanced Mathematical Concepts in Research Papers? by nirnamous
How about show this reference to chatgpt and ask again?
GufyTheLire t1_jc3evhf wrote
Reply to comment by RoboiosMut in Recommendations sources for Understanding Advanced Mathematical Concepts in Research Papers? by nirnamous
I've tried that once. Asked ChatGPT why L0, L1.. Ln norms, so seemingly different, were all named in a similar way. It correctly listed the norms' definitions and use cases, but failed to generalize the concept and made up some bullshit reason why they are named like that. Took me some time down the Wikipedia and Google rabbit hole to find out about Lp spaces and substitute different p values in the definition of p-norm to get the real reason
notgettingfined t1_jc33bc3 wrote
Reply to comment by grid_world in Image reconstruction by grid_world
I don’t under how that prevents learning from the individual images. I think you would need to explain the problem better. You could also add all the images together into channels so so you would have a 36x90x90input and then a 3x90x90 output
grid_world OP t1_jc32gs7 wrote
Reply to comment by notgettingfined in Image reconstruction by grid_world
I don’t want to mix up the individual information contained in 18 RGB images together in the hopes the the network learns the anticipated features out of them
notgettingfined t1_jc30vh5 wrote
Reply to Image reconstruction by grid_world
Why can’t you stitch together the 18 images into a single input?
suflaj t1_jc6n8v1 wrote
Reply to Image reconstruction by grid_world
Just apply an aggregation function on the 0th axis. This can be sum, mean, min, max, whatever. The best is sum, since your loss function will naturally regularise the weights to be smaller and it's the easiest to differentiate. This is in the case you know you have 18 images, for the scenario where you will have a variable amount of images, use mean. The rest are non-differentiable and might give you problems.
If you use sum, make sure you do gradient clipping so the gradients don't explode in the beginning.