itsnotlupus

itsnotlupus t1_jdj2xpr wrote

Meh. We see a few demos and all of the demos work all of the time, but that could easily be an optical illusion.

Yes, GPT-4 is probably hooked to subsystems that can parse an image, be it some revision of CLIP or whatever else, and yes it's going to work well enough some of the time, maybe even most of the time.

But maybe wait until actual non-corpo people have their hands on it and can assess how well it actually works, how often it fails, and whether anyone can actually trust it to do those things consistently.

1

itsnotlupus t1_jd5td54 wrote

It's not a user-facing product, it's a building block that would be useful to train music-oriented neural network, be they diffusers or other types of models.

It's probably going to take a little while before we see new models that leverage this library.

If you're looking for "stable diffusion but for music" right now, you could look at Riffusion (https://huggingface.co/riffusion/riffusion-model-v1)

2

itsnotlupus t1_j8k977o wrote

I just got access to the kitchen, and I'm a bit underwhelmed. No free conversation prompts, 3 very limited scenarios with a limited number of consecutive back and forth before getting thrown out and having to start from zero.
I had a conversation with a tennis ball obsessed with dog where almost every answer was something like "I don't know anything about that, but man, dogs are so cool, right?"
I did get it to admit cats were fun to play with too, albeit not as fun as dogs, shortly before being told the conversation was over.

I don't know if you've got access to more open-ended flavors of LaMDA, but for me this was hard to compare favorably to ChatGPT, warts and all.

1

itsnotlupus t1_j8ft0dq wrote

Well, people trust them today. They shouldn't, but they do. And it's going to get hilarious.

More seriously, we're going to learn collectively to flex a new muscle of "this AI may be super helpful, but it may also be bullshitting me." And odds are it'll be a bit of both in every answer.

Maybe those models are the inoculation we need to practice detecting bullshit online?

12

itsnotlupus t1_j2tbhzu wrote

Can you prune a pruned model? And then prune that again?

There's apparently no retraining needed here. Just loop over the matrices and shrink them (although it'd be nicer if there was a code repo to actually see that in action.)

I get that each successive pruning is going to make things increasingly worse, but I'm wondering if this might mean you can take an OPT-175B model and shrink it down in size to fit on commodity hardware like OPT-6.7B while still being closer in performance to the larger initial model than to the natively smaller model.

2