Viewing a single comment thread. View all comments

IntelArtiGen t1_ir88qh5 wrote

>While our internal testing suggest much of explicit and violent content can be filtered out, there still exists social biases and stereotypes which are challenging to detect and filter. We have decided not to release the Imagen Video model or its source code until these concerns are mitigated.

I think they'll never be mitigated and we'll have to wait for other people trying to reproduce the results and make them open-source.

173

Erosis OP t1_ir8cdlx wrote

It seems that Google is being very conservative with the release of their diffusion models compared to even Meta and OpenAI's closed-source approach.

Luckily, Stability AI seems to be working on a video generating diffusion model.

77

KeikakuAccelerator t1_ir990ti wrote

Meta (FAIR) has been very open-source.

13

Erosis OP t1_ir9kj9k wrote

I'm referring to their new Make-A-Video model, but I suppose they just announced that a few days ago. Hopefully they fully release that model.

13

gwern t1_irassaq wrote

They said they were considering releasing Make-A-Scene but never wound up doing so, even though it's probably not much better than the released SD model and there would seem to be fairly minimal marginal harm from a release. So I don't expect Make-A-Video to be released either, even if they say they might.

6

dexter89_kp t1_iragyoj wrote

That was trained on shutterstock data. They can’t release it

−1

gpahul t1_irbobh4 wrote

I noticed MS is also contributing in doc related AI research!

1

zaptrem t1_ir8ws0l wrote

Don’t worry, all we have to do to make the problem go away is end racism and all other biases in real life!

41

BalorNG t1_irdv27c wrote

Yea. The problem is never with the model - it is with the people. In a way, models trained on huge corpus of data is the most "democratic" way of representing reality - removing "biases" from it is castrating it. Those that are to exploit those biases needs to be dealt with on an individual basis.

3

IanMazgelis t1_ir9q097 wrote

This alone is why I have zero interest in proprietary diffusion algorithms. I want to make whatever I can imagine, not what shareholders tell I'm allowed to think about.

26

EmbarrassedHelp t1_ircm1je wrote

Its certainly harder to trust closed source implementations can do what they claim to do

1

brates09 t1_irhz7ml wrote

Are there examples of the recent big model work that haven’t been able to be replicated in terms of quality? Seems much more likely to attribute to conservatism of the companies rather than deception about the results.

1

101111010100 t1_ir9axqd wrote

Thank god humanity is still save. Once there are open-source versions, a lot of people will be harmed. /s

19

sam__izdat t1_irc0229 wrote

Any confluence with the interests of private capital is purely coincidental with our selfless quest for the betterment of man.

3

yaosio t1_irch8j4 wrote

It burns my bread that they are always worried about explicit scenes and violence. If they were producers for Django Unchained they would demand all the violence and bad words be removed.

2

ellioso t1_ir9ncfb wrote

It'll probably never be perfect but that doesn't mean it won't get released. The SafeSearch filter for google is really good.

2

cleverestx t1_irbc5ii wrote

Hopefully unstable diffusion gets a hold of this, and implements this technology, so people don't have to wait forever to be able to create stuff, as growups should be able to, without anything besides the most extreme legal restrictions in place.

2

Decent-Possible-9714 t1_irk8q2e wrote

what are the true technical limitations of their model (other than the obvious improve training/testing acc.)? they don't seem to explicitly detail it (meta explained numerous issues with their text-to-video model).

1