starstruckmon
starstruckmon t1_j6t4l60 wrote
Reply to comment by Flaky_Preparation_50 in Former Instagram Co-Founders Launch AI-Powered Personalized News App, Artifact by Flaky_Preparation_50
Good luck. I think there's a good market for such an offering.
starstruckmon t1_j6stlml wrote
Reply to Former Instagram Co-Founders Launch AI-Powered Personalized News App, Artifact by Flaky_Preparation_50
Just another recommendation system. What I'd like is a bot that crawls the web for all articles ( or even discussions ) of a single story, strips out all the narrative, opinion and fluff and just gives me an aggregated bullet point list.
Submitted by starstruckmon t3_10qhgmv in MachineLearning
starstruckmon t1_j6l1k5l wrote
Reply to Prompt engineering by im-so-stupid-lol
>take a human and show them 4 or 5 images of an animal they've never seen before they'll generally be able to draw it quite well
4-5 is actually enough to fine tune a pretrained SD model. Which is the correct comparison since we're already pretrained. Even if you ignore all the data upto that point in your life, even newborn brains are pretrained by evolution. They aren't initialised from random weights. Easier to notice this in other animals that can start walking right after birth.
starstruckmon t1_j6l12fe wrote
This is already true with photoshop. That's why source and chain of custody is important.
starstruckmon t1_j6l0a56 wrote
Reply to comment by sumane12 in What jobs will be one of the last remaining ones? by MrCensoredFace
Very few patients are going to be okay with robotic surgery that isn't supervised by a doctor. It doesn't even matter if it's technically better. Patients are just not going to trust it. Same with pilots. Even if it's fully auto-pilot they'll want someone to take control if something goes wrong. Trains are much easier to automate yet they still have an engineer/driver.
starstruckmon t1_j6kygds wrote
Reply to comment by _Just7_ in Chinese Search Giant Baidu to Launch ChatGPT-Style Bot by Buck-Nasty
I can't really speculate on that topic. It's currently an active area of research.
To be honest, this problem is so widely known that I hadn't considered finding sources to support the claim. Here is the best authoritative source I could quickly find
https://arxiv.org/abs/2012.15613
It may seem counter-intuitive to link to a paper that supposedly fixes this issue, but this is obviously the most likely scenario in which a paper would discuss it. Also, if you read it carefully, you'll see that while the authors managed to reduce the gap, it still persists.
starstruckmon t1_j6jw3kl wrote
Reply to comment by visarga in Chinese Search Giant Baidu to Launch ChatGPT-Style Bot by Buck-Nasty
It seems like you're talking about a model that has been trained in both languages. However, there are two issues with this. Firstly, the Chinese generally prefer to train models solely on Chinese data or with a limited amount of English data included. Secondly, multi-language models currently perform significantly worse compared to models that are trained on a single language.
starstruckmon t1_j6j4jxi wrote
Reply to comment by ihateshadylandlords in Parsel: A (De-)compositional Framework for Algorithmic Reasoning with Language Models - Stanford University Eric Zelikman et al - Beats prior code generation sota by over 75%! by Singularian2501
Even if the LLMs themselves don't become perfect at generating Parcel psudocode, having a compiler LM that can reliably convert Parcel ( or something simmilar ) to actual code would be a massive win. Imagine coding in natural language psudocode. A high-er level programming language.
starstruckmon t1_j6j05sx wrote
Reply to comment by FirstEbb2 in Chinese Search Giant Baidu to Launch ChatGPT-Style Bot by Buck-Nasty
At the beginning ( before you got to the downloading app part ), I thought you were starting something about Google. That's how bad Google's quality has become recently.
starstruckmon t1_j6izowe wrote
Reply to comment by Melancholy-Zebra in Chinese Search Giant Baidu to Launch ChatGPT-Style Bot by Buck-Nasty
I would be very surprised. They have technically speaking ( as per benchmarks ), one of the best text-to-image generators right now, yet the practical output is far below what we have in quality due to the limited dataset.
It would probably be even worse for text. Wikipedia, reddit, all the code forums like stackoverflow, documentations and manuals, vast majority of scientific papers. They'd be leaving so much out.
starstruckmon t1_j6i1ta1 wrote
Reply to comment by Primo2000 in Meta's chief AI scientist says "ChatGPT is not innovative". by ZaKodiak
Did you mix up Google and Meta there?
starstruckmon t1_j6erqgo wrote
Reply to comment by [deleted] in OpenAI has hired an army of contractors to make basic coding obsolete by Buck-Nasty
Who's calling GPT3 AGI?
starstruckmon t1_j6dmu5t wrote
Reply to comment by Kolinnor in My human irrationality is already taking over: as generative AI progresses, I've been growing ever more appreciative of human-made media by Yuli-Ban
I was going to do this before coming to the comments to see someone already had. Thank you.
starstruckmon t1_j6d3lsr wrote
Reply to comment by Maximum-Nectarine-13 in [D] MusicLM: Generating Music From Text by carlthome
I can guarantee the next paper out of this Google team is going to be a diffusion model ( instead of AudioLM ) conditioned on MuLan embeddings.
The strength of the Google model is the text understanding which is coming from the MuLan embeddings. While the strength of the work you highlighted is the quality from the diffusion model.
It's the obvious next step following the same path as Dalle1->Dalle2.
starstruckmon t1_j620ql2 wrote
Open up the examples and go listen to the dual conditioning. Humming/whistling/guitar/piano ( or anything really ) to any type of music.
starstruckmon OP t1_j501y7y wrote
Reply to comment by johnrachwan in [R] Massive Language Models Can Be Accurately Pruned in One-Shot by starstruckmon
From the paper
>One natural avenue for future work would be to investigate fine-tuning mechanisms for such large-scale models, which would allow further accuracy recovery. We conjecture that this should be possible, and that probably at least 80-90% sparsity can be achieved with progressive pruning and fine-tuning.
So, that comes next. Though I doubt the 80-90% guesstimate.
starstruckmon t1_j4uufbc wrote
You don't really need a separate extension, do you? Your bot can just be another user submitting the timestamps.
Though it would help if the extension developer provided a list of videos that are being watched by their users but has no timestamps yet, so your bot isn't spending time scraping though unpopular videos.
starstruckmon t1_j41dgsk wrote
Reply to comment by visarga in [D] Microsoft ChatGPT investment isn't about Bing but about Cortana by fintechSGNYC
There's no way for us to tell for certain, but since Google has used it for creativity oriented projects/papers like Dramatron, I don't think so. I feel the researchers would have said something instead of leading the whole world intentionally astray as everyone is now following Chinchilla's scaling laws.
Chinchilla isn't just a smaller model. It's adequately trained unlike GPT3 which is severely undertrained, so simmilar, if not exceeding ( as officially claimed ), capabilities isn't unexpected.
starstruckmon t1_j3x9azk wrote
Reply to comment by slashd in [D] Microsoft ChatGPT investment isn't about Bing but about Cortana by fintechSGNYC
I'm not sure this really is a winner takes all market but maybe. Good point.
starstruckmon t1_j3x6dzj wrote
Reply to comment by m98789 in [D] Microsoft ChatGPT investment isn't about Bing but about Cortana by fintechSGNYC
Fair enough.
starstruckmon t1_j3x4bsr wrote
Reply to comment by erelim in [D] Microsoft ChatGPT investment isn't about Bing but about Cortana by fintechSGNYC
How is Google behind OpenAI? Chinchilla has simmilar performance as GPT3 yet is much cheaper to run since it has less than half the parameters.
starstruckmon t1_j3wxccu wrote
Reply to comment by yaosio in [D] Microsoft ChatGPT investment isn't about Bing but about Cortana by fintechSGNYC
True and that's probably the reason. But still, they have a ML/AI division. Why not have them just train Megatron to convergence and leapfrog GPT3? I'll never understand how these companies make decisions honestly.
starstruckmon t1_j3wvdqt wrote
Reply to comment by m98789 in [D] Microsoft ChatGPT investment isn't about Bing but about Cortana by fintechSGNYC
>I think you may be underestimating the compute cost. It’s about $6M of compute (A100 servers) to train a GPT-3 level model from scratch. So with a billion dollars, that’s about 166 models.
I was actually overestimating the cost to train. I honestly don't see how these numbers don't further demonstrate my point. Even if it cost a whole billion ( that's a lot of experimental models ), that's still 10 times less than what they're paying.
>Considering experimentation, scaling upgrades, etc., that money will go quickly. Additionally, the cost to host the model to perform inference at scale is also very expensive. So it may be the case that the $10B investment isn’t all cash, but maybe partially paid in Azure compute credits. Considering they are already running on Azure.
I actually expect every last penny to go into the company. They definitely aren't buying anyone's shares ( other than maybe a partial amount of employee's vested shares ; this is not the bulk ). It's mostly for new shares created. But $10B for ~50% still gives you a pre-money valuation of ~10B. That's a lot.
starstruckmon t1_j6tchwa wrote
Reply to comment by Flaky_Preparation_50 in Former Instagram Co-Founders Launch AI-Powered Personalized News App, Artifact by Flaky_Preparation_50
To be perfectly honest, there really isn't much there to comment on currently.