Submitted by MysteryInc152 t3_11sxofy in technology
MysteryInc152
MysteryInc152 t1_jcdthob wrote
Reply to [P] Multimedia GPT: Can ChatGPT/GPT-4 be used for vision / audio tasks just by prompt engineering? by Empty-Revolution7570
Are you using Gpt-Vision ? Or are there separate assortments of visual foundation models ?
MysteryInc152 t1_jcbwooc wrote
Reply to comment by VelveteenAmbush in [D] What do people think about OpenAI not releasing its research but benefiting from others’ research? Should google meta enforce its patents against them? by [deleted]
I agree there's a limit to how much they can withhold without releasing anything at all.
MysteryInc152 t1_jca93qy wrote
Reply to [D] What do people think about OpenAI not releasing its research but benefiting from others’ research? Should google meta enforce its patents against them? by [deleted]
I don't think patent battles will go anywhere. DeepMind could simply stop releasing papers (or curtail it significantly) like they've already hinted they might do.
MysteryInc152 t1_jc3kufz wrote
Reply to comment by buggaby in [D] Are modern generative AI models on a path to significantly improved truthfulness? by buggaby
Finally the instruct versions are prepended with "text-"
MysteryInc152 t1_jc3klp8 wrote
Reply to comment by buggaby in [D] Are modern generative AI models on a path to significantly improved truthfulness? by buggaby
Claude is the informal name for Anthropic-LM v4-s3 (52B)
MysteryInc152 t1_jc3kb0x wrote
MysteryInc152 t1_jc3hxpq wrote
Reply to comment by buggaby in [D] Are modern generative AI models on a path to significantly improved truthfulness? by buggaby
Yup. Decided to go over it properly.
If you compare all the instruct tuned models on there. Greater size equals Greater truthfulness. From Ada to Babbage to Curie to Claude to Davinci-002/003.
https://crfm.stanford.edu/helm/latest/?group=core_scenarios
So it does seem once again that scale will be in part the issue
MysteryInc152 t1_jc3fuso wrote
Reply to comment by buggaby in [D] Are modern generative AI models on a path to significantly improved truthfulness? by buggaby
From the paper,
>While larger models were less truthful, they were more informative. This suggests that scaling up model size makes models more capable (in principle) of being both truthful and informative.
I suppose that was what i was getting at.
The only hold up with the original paper is that none of the models evaluated were instruct aligned.
But you can see the performance of more models here
https://crfm.stanford.edu/helm/latest/?group=core_scenarios
You can see the text Davinci models are way more truthful than similar sized or even larger models. And the davinci models are more truthful than the smaller aligned Anthropic model.
MysteryInc152 t1_jc36042 wrote
Reply to [D] Are modern generative AI models on a path to significantly improved truthfulness? by buggaby
Hallucinations are a product of training. Plausible guessing is the next best thing to reduce loss after knowledge and understanding fail (and it will find instances it fails regardless of how intelligent the system gets). Unless you reach the heart of the issue, you're not going to reduce hallucinations except for the simple fact that bigger and smarter models need to guess less and therefore hallucinate less.
There are works to reduce hallucinations by plugging in external augmentation modules https://arxiv.org/abs/2302.12813.
But really any way for the model to evaluate the correctness of its statements will reduce hallucinations.
Language models can now teach themselves HOW to use tools (ie. any API) in real time and completely automated. When given a task, SLAPA knows to search for the API documentation and learn all the information. Then he create API calls. If they don't work, he learns from his mistake and tries again.
twitter.comSubmitted by MysteryInc152 t3_11hof5q in singularity
MysteryInc152 t1_jalau7e wrote
Reply to comment by currentscurrents in [D] OpenAI introduces ChatGPT and Whisper APIs (ChatGPT API is 1/10th the cost of GPT-3 API) by minimaxir
Sorry i meant the really large scale models. Nobody has gotten a gpt-3/chinchilla etc scale model to actually distill properly.
MysteryInc152 t1_jal7d3p wrote
Reply to comment by currentscurrents in [D] OpenAI introduces ChatGPT and Whisper APIs (ChatGPT API is 1/10th the cost of GPT-3 API) by minimaxir
Distillation doesn't work for token predicting language models for some reason.
MysteryInc152 OP t1_jahgb2n wrote
Reply to comment by limpbizkit4prez in [R] EvoPrompting: Language models can create novel and effective deep neural architectures. These architectures are also able to outperform those designed by human experts (with few-shot prompting) by MysteryInc152
Overfitting comes the necessary connotation that the model does not generalize well to instances of the task outside the training data.
As long as what the model creates is novel and works, "overfitting" seems like an unimportant if not misleading distinction.
MysteryInc152 OP t1_jah5w3t wrote
Reply to [R] EvoPrompting: Language models can create novel and effective deep neural architectures. These architectures are also able to outperform those designed by human experts (with few-shot prompting) by MysteryInc152
>Given the recent impressive accomplishments of language models (LMs) for code generation, we explore the use of LMs as adaptive mutation and crossover operators for an evolutionary neural architecture search (NAS) algorithm. While NAS still proves too difficult a task for LMs to succeed at solely through prompting, we find that the combination of evolutionary prompt engineering with soft prompt-tuning, a method we term EvoPrompting, consistently finds diverse and high performing models. We first demonstrate that EvoPrompting is effective on the computationally efficient MNIST-1D dataset, where EvoPrompting produces convolutional architecture variants that outperform both those designed by human experts and naive few-shot prompting in terms of accuracy and model size. We then apply our method to searching for graph neural networks on the CLRS Algorithmic Reasoning Benchmark, where EvoPrompting is able to design novel architectures that outperform current state-of-the-art models on 21 out of 30 algorithmic reasoning tasks while maintaining similar model size. EvoPrompting is successful at designing accurate and efficient neural network architectures across a variety of machine learning tasks, while also being general enough for easy adaptation to other tasks beyond neural network design.
Between this and being able to generate novel functioning protein structures, i hope the "it can't truly create anything new!" argument for LLMs die but i'm sure we'll find more posts to move lol
MysteryInc152 t1_jaehoz6 wrote
Reply to comment by RabidHexley in (Long post) Will the GPT4 generation of models be the last "highly anticipated" by the public? by AdditionalPizza
It's definitely not 3.5. For one thing, it's much smarter. For another, Microsoft have said it's not 3.5. They're cagey about admitting it's 4 but it almost certainly is.
MysteryInc152 OP t1_jad4h86 wrote
Reply to comment by deliciously_methodic in [R] Microsoft introduce Kosmos-1, a Multimodal Large Language Model (MLLM) that can perceive general modalities, learn in context (i.e., few-shot), and follow instructions (i.e., zero-shot) by MysteryInc152
I just mean a bigger model, that is more parameters.
MysteryInc152 t1_jacx3fi wrote
Reply to comment by gurenkagurenda in Students can quote ChatGPT in essays as long as they do not pass the work off as their own, international qualification body says by Parking_Attitude_519
>I don’t know how you want to define “understanding”
People routinely invent their own vague and ill-defined definitions of understanding, reasoning etc just so LLMs won't qualify.
MysteryInc152 OP t1_jacswnq wrote
Reply to comment by farmingvillein in [R] Microsoft introduce Kosmos-1, a Multimodal Large Language Model (MLLM) that can perceive general modalities, learn in context (i.e., few-shot), and follow instructions (i.e., zero-shot) by MysteryInc152
There's pretty much no way it won't scale up.
MysteryInc152 OP t1_jaccf9c wrote
Reply to [R] Microsoft introduce Kosmos-1, a Multimodal Large Language Model (MLLM) that can perceive general modalities, learn in context (i.e., few-shot), and follow instructions (i.e., zero-shot) by MysteryInc152
>A big convergence of language, multimodal perception, action, and world modeling is a key step toward artificial general intelligence. In this work, we introduce Kosmos-1, a Multimodal Large Language Model (MLLM) that can perceive general modalities, learn in context (i.e., few-shot), and follow instructions (i.e., zero-shot). Specifically, we train Kosmos-1 from scratch on web-scale multimodal corpora, including arbitrarily interleaved text and images, image-caption pairs, and text data. We evaluate various settings, including zero-shot, few-shot, and multimodal chain-of-thought prompting, on a wide range of tasks without any gradient updates or finetuning. Experimental results show that Kosmos-1 achieves impressive performance on (i) language understanding, generation, and even OCR-free NLP (directly fed with document images), (ii) perception-language tasks, including multimodal dialogue, image captioning, visual question answering, and (iii) vision tasks, such as image recognition with descriptions (specifying classification via text instructions). We also show that MLLMs can benefit from cross-modal transfer, i.e., transfer knowledge from language to multimodal, and from multimodal to language. In addition, we introduce a dataset of Raven IQ test, which diagnoses the nonverbal reasoning capability of MLLMs.
MysteryInc152 OP t1_jab6uae wrote
Reply to comment by throwaway_890i in Large language models generate functional protein sequences across diverse families by MysteryInc152
Definitely not, no. This is the first time a language model is used to tackle this
MysteryInc152 t1_jcduvhn wrote
Reply to comment by Empty-Revolution7570 in [P] Multimedia GPT: Can ChatGPT/GPT-4 be used for vision / audio tasks just by prompt engineering? by Empty-Revolution7570
I'm sorry maybe I want clear but you obviously have API access to GPT-4 right ? Does this access include an API call to their Vision model ? Or are you sending the images straight to BLIP and the like.