shmoculus

shmoculus t1_jcl1h18 wrote

Reply to comment by flyblackbox in Those who know... by Destiny_Knight

I share this view. Another thing is that these single models don't scale, you'll want them to access othe models, different data sources etc, for that you need permission less ways to transact value on demand, which is the entire premise of crypto. Example is your llm need to access recent data on X to make a decision, access to data for X is via paid subscription, not gonna work, need way to access paid data ad hoc without credit card anonymously, crypto smart contracts are the way

2

shmoculus t1_jcfcpd5 wrote

I think generative ai will become vr / ar's killer app. Once image generation is fast enough to be percieved as reality, you can explore endless worlds, stories and adventures.

Interacting with artificial agents is going to be a game changer in virtual spaces

22

shmoculus t1_j9yd3ep wrote

I'm noticing competitve edge doesn't last as long as it used to and it's hard to predict how long a business model will be relevant e.g.

5 years ago, startups were offering customised chatbots, now ChatGPT style variants will replace those

NovelAI was apparently the best at anime image generation and its model got leaked and community trained models reduced the need for that kind of service.

elevenlabs has the best voice generator, if an open source model became available at similar quality, why would most people pay elevenlabs?

It seems a business model based purely on tech. advantage is risky because of security risks and a motivated open source community. e.g if someone leaks your model, that's a huge amount of investment and competitive edge lost overnight.

I think people would pay for convenience, as running these models is challenging for lay people e.g. midjourney

1

shmoculus t1_j7tfb1n wrote

I think Google underestimated the competition. They prob thought that since they couldn't release a perfect chatbot that others were well behind. Well OpenAI caught them with their pants down and Microsoft was quick to cement the advantage.

Google is good at innovating but terrible at delivering consumer products. MS is decent at delivering consumer products but terrible at innovating. Well they out sourced the innovation this time

2

shmoculus t1_j7ilnqr wrote

Google search, Google docs, Andorid, Gmail. Almost everything else is a pain in the ass to use. TensorFlow vs Pytorch, Angular vs React and maybe even Kubernetes vs something I'm not aware of. Remember Google Plus, Google Glass all of their labs products they abandonded? They're not great at delivering consumer products. Not to mention their search, youtube and maps etc is just of full ad driven bloat now

1

shmoculus t1_j4p9la8 wrote

Reply to comment by Antok0123 in Question about AI art. by cloudrunner69

For example Stable Diffusion is based on the LAION datasets which include billions of images scraped from the internet. So Phillipino culture is likely underrepresented on the general internet and the models don't have a good representation of them.

People take the trained models and add stuff that wasn't in the training set, e.g. you can train on images of 1600s philippines to get what you want that is currently missing.

Have a look here for some custom models, people have added styles, concepts, people etc: https://civitai.com/, you could easily make one that does historical periods from all over the world. I'm sure people would love it.

6

shmoculus t1_j36xwy5 wrote

You know, it's always the details that get in the way, namely big data to train the models (scraping, storage, cleaning), big compute to train the models (read $$$,$$$), just endless boiler plate engineering work to get products up and running, then ongoing costs, challenges in running large models locally, continuous improvement etc.

Now having said all that, you can participate in LAION's OpenAssistant, have at it friend :)

7

shmoculus t1_j290fym wrote

I think it's easier for people to understand AGI as a reasoning machine, reason is not necessarily tied to being conscious / self-awareness (though some self awareness helps in acting in the world so will likely be implicitly learned)

1

shmoculus t1_j28yfux wrote

I kind of see what you are getting at, and it could be the case with exponential improvements in methods/research that we see more discoveries in one year than all the previous at some point but I don't think we're there yet.

The progression has been linear in my view:

  1. Efficient image classification (CNNs)

  2. object detection / segmentation / pix2pix / basic img2text models (RCNNs, Unet, GANs)

  3. Deep reinforcement learning (DQN, PPO, MCTS)

  4. Attention networks (transformers and language modelling)

  5. Basic question / answer and reasoning models

  6. Low quality txt2img models (e.g. DALL-E 1)

  7. High quality txt2img models (e.g. DALL-E 2, stable diffusion)

  8. Multimodal modals (image understading etc) <- we are here

  9. Already happening video2video models, text2mesh / point cloud

  10. Expect low, then high quality multimodal generation models e.g. txt2video + music

  11. Expect improved text understanding, general chat behaviour, ie large step ups in chatbot usefulness inclution ability to take actions (this part is already underway)

  12. Expect some kind of attention based method for reading and writing to storage (i.e memory) and possibly online learning / continuous improvement

13 . More incrementally interesting stuff :)

5