Submitted by Erosis t3_xws0p1 in MachineLearning
IntelArtiGen t1_ir88qh5 wrote
>While our internal testing suggest much of explicit and violent content can be filtered out, there still exists social biases and stereotypes which are challenging to detect and filter. We have decided not to release the Imagen Video model or its source code until these concerns are mitigated.
I think they'll never be mitigated and we'll have to wait for other people trying to reproduce the results and make them open-source.
Erosis OP t1_ir8cdlx wrote
It seems that Google is being very conservative with the release of their diffusion models compared to even Meta and OpenAI's closed-source approach.
Luckily, Stability AI seems to be working on a video generating diffusion model.
KeikakuAccelerator t1_ir990ti wrote
Meta (FAIR) has been very open-source.
Erosis OP t1_ir9kj9k wrote
I'm referring to their new Make-A-Video model, but I suppose they just announced that a few days ago. Hopefully they fully release that model.
gwern t1_irassaq wrote
They said they were considering releasing Make-A-Scene but never wound up doing so, even though it's probably not much better than the released SD model and there would seem to be fairly minimal marginal harm from a release. So I don't expect Make-A-Video to be released either, even if they say they might.
dexter89_kp t1_iragyoj wrote
That was trained on shutterstock data. They can’t release it
gpahul t1_irbobh4 wrote
I noticed MS is also contributing in doc related AI research!
zaptrem t1_ir8ws0l wrote
Don’t worry, all we have to do to make the problem go away is end racism and all other biases in real life!
BalorNG t1_irdv27c wrote
Yea. The problem is never with the model - it is with the people. In a way, models trained on huge corpus of data is the most "democratic" way of representing reality - removing "biases" from it is castrating it. Those that are to exploit those biases needs to be dealt with on an individual basis.
IanMazgelis t1_ir9q097 wrote
This alone is why I have zero interest in proprietary diffusion algorithms. I want to make whatever I can imagine, not what shareholders tell I'm allowed to think about.
EmbarrassedHelp t1_ircm1je wrote
Its certainly harder to trust closed source implementations can do what they claim to do
brates09 t1_irhz7ml wrote
Are there examples of the recent big model work that haven’t been able to be replicated in terms of quality? Seems much more likely to attribute to conservatism of the companies rather than deception about the results.
101111010100 t1_ir9axqd wrote
Thank god humanity is still save. Once there are open-source versions, a lot of people will be harmed. /s
sam__izdat t1_irc0229 wrote
Any confluence with the interests of private capital is purely coincidental with our selfless quest for the betterment of man.
yaosio t1_irch8j4 wrote
It burns my bread that they are always worried about explicit scenes and violence. If they were producers for Django Unchained they would demand all the violence and bad words be removed.
ellioso t1_ir9ncfb wrote
It'll probably never be perfect but that doesn't mean it won't get released. The SafeSearch filter for google is really good.
cleverestx t1_irbc5ii wrote
Hopefully unstable diffusion gets a hold of this, and implements this technology, so people don't have to wait forever to be able to create stuff, as growups should be able to, without anything besides the most extreme legal restrictions in place.
Decent-Possible-9714 t1_irk8q2e wrote
what are the true technical limitations of their model (other than the obvious improve training/testing acc.)? they don't seem to explicitly detail it (meta explained numerous issues with their text-to-video model).
Viewing a single comment thread. View all comments