Viewing a single comment thread. View all comments

FirstOrderCat t1_itad0zs wrote

Reply to comment by AsthmaBeyondBorders in U-PaLM 540B by xutw21

>The problem is you don't know what emergent skills are yet to be found because we didn't scale enough.

Yes, and you don't know if such skills will be found and we hit the wall or not yet.

1

AsthmaBeyondBorders t1_itad4q2 wrote

There is a very old solution to finding that out. It is to scale and check instead of guessing

5

FirstOrderCat t1_itaeypy wrote

this race maybe over.

On the graph guy is proud of getting 2 points in some synthetic benchmark, while spending 4 millions TPUv4 hours = $12M.

At the same time we hear that Google cuts expenses and considering layoffs, and LLM part of Google Research will be the first in the line, because they don't provide much value in Ads/Search business.

1

AsthmaBeyondBorders t1_itagbfz wrote

This model had up to 21% gains in some benchmarks, as you can see there are many benchmarks. You may notice this model is still 540B just like the older one, so this isn't about scale it is about a different model which can be as good and better than the previous ones while cheaper to train.

You seem to know a lot about Google's internal decisions and strategies as of today, good for you, I can't discuss stuff I have absolutely no idea about and clearly you have insider information about where google is going and what they are doing, that's real nice.

3

FirstOrderCat t1_itahazn wrote

>This model had up to 21% gains in some benchmarks, as you can see there are many benchmarks

Meaning they received less than 2 points in many others..

> it is about a different model which can be as good and better than the previous ones while cheaper to train.

Model is the same, they changed training procedure.

> You seem to know a lot about Google's internal decisions and strategies as of today

This is public information.

2

justowen4 t1_itau79p wrote

Epic commenting you two. The winner is….. AsthmaBeyondBorders !

2