Nmanga90

Nmanga90 t1_j8ar7ic wrote

Not right now it won’t. We already know of ways to improve AI, but we don’t have data to allow an AI to improve itself. The only way for that would be generative design, which is by nature very wasteful and slow. Once it gets to a certain point yes, but as of right now we are far (relatively) from that

1

Nmanga90 t1_j87t99b wrote

It is I promise. We have barely broken the surface of what AI can do with human designs. Right now the main limitations are compute power and data. Any explorations into alternative architectures and whatnot come with a massive opportunity cost because of these.

OpenAI alone has probably spent a billion on compute up to now. Insanity

I don’t think you guys understand, but every single week we’re improving leaps and bounds with minor tweaks and modifications to existing architecture. It would be extremely inefficient to allow the AI to try and improve itself when we have almost guaranteed improvement by humans that is only limited by how much GPU we can muster

3

Nmanga90 t1_j78du1g wrote

Reply to comment by Akimbo333 in Infinite police by crap_punchline

Just out of curiosity, what is your education on the subject? I find it kind of strange or I guess inconsistent that you’re talking about multimodal LLMs and their necessity, but don’t know about OPT, InstructGPT, or why an Instruct model would be better than a predictive model

1

Nmanga90 t1_j784rkz wrote

Reply to comment by Akimbo333 in Infinite police by crap_punchline

What exactly don’t you understand?

Following instructions makes it better because these models are by nature predictive. They don’t understand what you are saying, and are created to predict the next text after the input. By nature, the models basically have an implicit prompt that says “what follows this input:”. This is much less useful than following instructions, because in the real world, there is less money/productivity to be gained by predicting the next text sequence, and more to be gained by completing tasks that you ask it to.

1

Nmanga90 t1_j77zkpf wrote

Reply to comment by Akimbo333 in Infinite police by crap_punchline

InstructGPT is GPT-3 fine tuned to follow instructions, and is now the flagship GPT3, and the newest davinci model is instructGPT. ChatGPT is based on instructGPT and further fine tuned for dialog.

1

Nmanga90 t1_j3xxwj6 wrote

Locally will not cut it unless you have a high performance computer with lab grade GPUs for inference. The reason the AI models are so expensive to use is because they are actually pretty expensive to run. They are running probably 2 parallel versions of the model on a single a100, and have likely duplicated this architecture 10,000 times. And an a100 is 10 grand used, 20 grand new. You can also rent them out for about $2 per minute.

1

Nmanga90 t1_j3xxi44 wrote

OpenAI is not going to close shop any time soon. Not sure if you know this, but Microsoft has been making huge investments into them, and has licensing rights to the GPT models. So Microsoft is pretty much the one who is serving the APIs, and they are right now looking into making another 10-billion-dollar investment into OpenAI.

1