currentscurrents
currentscurrents t1_j865is4 wrote
There's some mutually assured destruction going on here. Microsoft/OpenAI also own patents that cover Google's products. If Google sued them over Transformers, they would sue right back for something else.
currentscurrents t1_j85rpol wrote
Reply to comment by goj-145 in [D] Is it legal to use images or videos with copyright to train a model? by Tlaloc-Es
They use the open LAION 50B dataset, everybody knows what's in there.
Still, some preprocessing and deduplication would have been a good idea just for output quality.
currentscurrents t1_j7y4073 wrote
Reply to comment by [deleted] in [D] Critique of statistics research from machine learning perspectives (and vice versa)? by fromnighttilldawn
>Right now basically all progress is with large models,
You mean all progress... in machine learning. A lot of scientific fields necessarily must make do with a smaller number of data points.
You can't test a new drug on a million people, especially in early phase trials. Even outside of medicine, you may have very few samples if you're studying a rare phenomena.
Statistics gives you tools to make limited conclusions from small samples, and also measure how meaningful those conclusions actually are.
currentscurrents t1_j7xv6j3 wrote
Reply to comment by [deleted] in [D] Critique of statistics research from machine learning perspectives (and vice versa)? by fromnighttilldawn
Stats is tremendously useful, especially when your dataset is small by ML standards. Basically every scientific paper relies on statistics to tell you whether or not their result is meaningful.
ML is great when you have millions of data points, but when you only have a hundred it's not going to help you.
currentscurrents t1_j7wuonk wrote
Reply to comment by big_ol_tender in [D]Image Recognition ability of machine learning in financial markets questions by Ready-Acanthaceae970
Agreed, financial markets have built-in protections against this kind of analysis. If it works, everyone else would do it, and the more people do it the less any of them benefit from it.
The only way to beat the market consistently is to have a source of information nobody else has access to, or at least hasn't discovered yet.
currentscurrents t1_j7wk84r wrote
Reply to comment by andreichiffa in [D] Are there emergent abilities of image models? by These-Assignment-936
While those are on the same topic, they're very different papers. The Anthropic paper spends most of its time going on about safety/bias/toxicity, while the Google paper is focused on more useful things like the technical abilities of the models.
currentscurrents t1_j7wf3u0 wrote
>What is the standard modeling approach to these kinds of problems?
The standard approach is reinforcement learning. It works, but it's not very sample-efficient and takes many iterations to train.
LLMs are probably so good at this because of their strong meta-learning abilities; during the process of pretraining they not only learn the task but also learn good strategies for learning new tasks.
This has some really interesting implications. Pretraining seems to drastically improve sample efficiency even if the pretraining was on a very different task. Maybe we could pretrain on a very large amount of synthetic, generated data before doing our real training on our finitely-sized real datasets.
currentscurrents t1_j7sri62 wrote
Reply to comment by katadh in [Discussion] Cognitive science inspired AI research by theanswerisnt42
SNN-ANN conversion is kludge - not only do you have to train an ANN first, it means your SNN is incapable of learning anything new.
Surrogate gradients are better! But they're still non-local and require backwards passes, which means you're missing out on the massive parallelization you could achieve with local learning rules on the right hardware.
Local learning is the dream, and would have benefits for ANNs too: you could train a single giant model distributed across an entire datacenter or even multiple datacenters over the internet. Quadrillion-parameter models would be technically feasible - I don't know what happens at that scale, but I'd sure love to find out.
currentscurrents t1_j7q8q5v wrote
Reply to comment by wintermute93 in [Discussion] Cognitive science inspired AI research by theanswerisnt42
So far nobody's figured out a good way to train them.
You can't easily do backprop, but you wouldn't want to anyway - the goal of SNNs is to run on ultra-low-power analog computers. For this you need local learning, where neurons can learn by communicating only with adjacent neurons. There's some ideas (forward-forward learning, predictive coding, etc) but so far nothing is as good as backprop.
There's a bit of a chicken-and-egg problem too. Without a good way to train SNNs, there's little interest in the specialized hardware - and without the hardware, there's little interest in good ways to train them. You can emulate them on regular computers but that removes all their benefits.
currentscurrents OP t1_j7mx9qp wrote
Reply to comment by theRIAA in [N] Microsoft announces new "next-generation" LLM, will be integrated with Bing and Edge by currentscurrents
That's a no from me dawg.
I'll wait my turn, I'm not installing their app.
currentscurrents OP t1_j7m04ot wrote
Reply to comment by buzzbuzzimafuzz in [N] Microsoft announces new "next-generation" LLM, will be integrated with Bing and Edge by currentscurrents
Meh, I think the safety concerns are overblown. It's really more of bad PR for Microsoft than an actual threat.
You can already find out how to make drugs, build a bomb, etc from the internet. The Anarchist Cookbook has been well-known for decades and you can find a pdf with a simple google search.
currentscurrents t1_j7j7dgu wrote
Reply to comment by PredictorX1 in Wouldn’t it be a good idea to bring a more energy efficient language into the ML world to reduce the insane costs a bit?[D] by thedarklord176
Call me when logistic regression can generate a realistic detailed digital painting by greg rutkowski.
currentscurrents t1_j7j78vc wrote
Reply to comment by thedarklord176 in Wouldn’t it be a good idea to bring a more energy efficient language into the ML world to reduce the insane costs a bit?[D] by thedarklord176
All the computation is happening on the GPU. Python is just making a bunch of calls to the GPU drivers.
Researchers spend a lot of time making neural networks as fast as possible. If switching to another language would have given a substantial speed boost, they would have done it already.
currentscurrents t1_j7iy068 wrote
Reply to comment by [deleted] in [N] Getty Images sues AI art generator Stable Diffusion in the US for copyright infringement by Wiskkey
The exception Google Images got is pretty narrow and only applies to their role as a search engine. Fair use is complex, depends on a lot of case law, and involves balancing several factors.
One of the factors is "whether your use deprives the copyright owner of income or undermines a new or potential market for the copyrighted work." Google Image thumbnails clearly don't compete with the original work, but generative AI arguably does - the fact that it could automate art production is one of the coolest things about it.
That said, this is only one of several factors, so it's not a slam dunk for Getty either. The most important factor is how much you borrow from the original work. AI image generators borrow only abstract concepts like style, while Google was reproducing thumbnails of entire works.
Anybody who thinks they know how the courts will rule on this is lying to themselves.
currentscurrents t1_j7iwvii wrote
Reply to comment by FyreMael in [N] Getty Images sues AI art generator Stable Diffusion in the US for copyright infringement by Wiskkey
> Also, LAION gathered the images via crowdsourcing. I participated.
I don't think the data collection methodology is really relevant. However the dataset was gathered, there are certainly ways to use it that would violate copyright. You couldn't print it in a book for example.
The important question is if training a generative AI on copyrighted data is a violation of copyright. US copyright law doesn't address this because AI didn't exist when it was written. It will be up to the courts to decide how this new application interacts with the law.
currentscurrents t1_j7ivm0a wrote
Reply to comment by trias10 in [N] Getty Images sues AI art generator Stable Diffusion in the US for copyright infringement by Wiskkey
>the only thing I have seen is cheating on homeworks and exams, faking legal documents, and serving as a dungeon master for D&D. The last one is kind of cool, but the first two are illegal.
Well that's just cherry-picking. LLMs could do very socially-good things like act as an oracle for all internet knowledge or automate millions of jobs. (assuming they can get the accuracy issues worked out - which there are tons of researchers trying to do, some of whom are even on this sub)
By far the most promising use is allowing computers to understand and express complex ideas in plain english. We're already seeing uses of this, for example text-to-image generators use a language model to understand prompts and guide the generation process. Or how Github Copilit can turn instructions from english into implementations in code.
I expect we'll see them applied to many more applications in the years to come, especially once desktop computers get fast enough to run them locally.
>starts playing by the same rules as everyone else in the industry.
Everyone else in the industry is also training on copyrighted data, because there is no source of uncopyrighted data big enough to train these models.
Also, your brain is updating its weights based on the copyrighted data in my comment right now, and that doesn't violate my copyright. Why should AI be any different?
currentscurrents t1_j7ipcip wrote
Reply to comment by trias10 in [N] Getty Images sues AI art generator Stable Diffusion in the US for copyright infringement by Wiskkey
OpenAI is doing a good thing. They've found a new and awesome way to use data from the open web, and they deserve their reward.
Getty's business model is outdated now, and the legal system shouldn't protect old industries from new inventions. Why search for a stock image that sorta kinda looks like what you want, when you could generate one that matches your exact specifications for free?
currentscurrents t1_j7ioshb wrote
Reply to comment by _poisonedrationality in [N] Getty Images sues AI art generator Stable Diffusion in the US for copyright infringement by Wiskkey
> Besides, it's not clear to me whether these AI tools be used to benefit humanity as a whole
Of course they benefit humanity as a whole.
- Language models allow computers to understand complex ideas expressed in plain english.
- Automating art production will make custom art/comics/movies cheap and readily available.
- ChatGPT-style AIs (if they can fix hallucination/accuracy problems) give you an oracle with all the knowledge of the internet.
- They're getting less hype right now, but there's big advances in computer vision (CNNs/Vision Transformers) that are revolutionizing robotics and image processing.
>I really hope this case sets ome decent precedents about how AI developers can use data they did not create.
You didn't create the data you used to train your brain, much of which was copyrighted. I see no reason why we should put that restriction on people trying to create artificial brains.
currentscurrents t1_j7innd5 wrote
Reply to comment by VeritaSimulacra in [N] Getty Images sues AI art generator Stable Diffusion in the US for copyright infringement by Wiskkey
Getty is just the test case for the question of copyright and AI.
If you can't train models on copyrighted data this means that they can't learn information from the web outside of specific openly-licensed websites like Wikipedia. This would sharply limit their usefulness. It also seems distinctly unfair, since copyright is only supposed to protect the specific arrangement of words or pixels, not the information they contain or the artistic style they're in.
The big tech companies can afford to license content from Getty, but us little guys can't. If they win it will effectively kill open-source AI.
currentscurrents t1_j6ycn5m wrote
Reply to LibreOffice 7.5 released by jlpcsl
Huh, I haven't used it in years but it's great to see them still chugging along!
currentscurrents t1_j6tn84b wrote
Reply to comment by Soft-Material3294 in [R] SETI finds eight potential alien signals with ML by logTom
Very high.
The standard of evidence required for aliens is also very high. It's not enough to have no known natural explanation; there's lots of natural phenomena we don't know about yet. It must affirmatively and unavoidably be artificial even after many follow-up observations.
currentscurrents t1_j6qgw2g wrote
Reply to comment by Square_Tea4916 in [OC] Manchester United Income and Expenses Breakdown of their 2022 Annual Report by Square_Tea4916
> they make their money on the sale of the franchise.
What does this mean? Are they planning to sell the business someday, and they will make their money back when they do so?
If the business is a net loss except when you sell it to someone else, why is anyone willing to buy it? EDIT: I guess people were willing to buy NFTs.
currentscurrents t1_j6pfewg wrote
Reply to comment by ialghamdi1 in [OC] McDonald's 2022 income statement (they just released their earnings) by IncomeStatementGuy
Those are overhyped. They haven't automated the actual cooking - just replaced the cashier with a touchscreen. Fully automated restaurants remain a dream.
currentscurrents t1_j6pf3yq wrote
Reply to comment by IncomeStatementGuy in [OC] McDonald's 2022 income statement (they just released their earnings) by IncomeStatementGuy
Of course, that's not the same as the actual profit margin of the restaurant - that's just what McDonald's corporate makes from the franchise vs the costs of what they provide to the franchise.
currentscurrents t1_j86gori wrote
Reply to The Inference Cost Of Search Disruption – Large Language Model Cost Analysis [D] by norcalnatv
In the long run, I think this is something that will be solved with more specialized architectures for running neural networks. TPUs and Tensor Cores are great first steps, but the Von Neumann architecture is holding us back.
Tensor Cores are very fast. But since the Von Neumann architecture has separate compute and memory connected by a bus, the entire network has to travel through the memory bus for every step of training or inference. The overwhelming majority of time is spent waiting on this:
>200 cycles (global memory) + 34 cycles (shared memory) + 1 cycle (Tensor Core) = 235 cycles.
A specialized architecture that physically implements neurons on silicon would no longer have this bottleneck. Since each neuron would be directly connected to the memory it needs (weights, data from previous layer) the entire network could run in parallel regardless of size. You could do inference as fast as you could shovel data through the network.