DeepGamingAI

DeepGamingAI t1_j7jq9jk wrote

To me all AI debate these days are just a regurgitation of "glass half full or half empty" discussions. Yes, LLMs are far more intelligent than anyone anticipated them to be by this point in time, and no they aren't general intelligence. The constant back and forth between these two groups can essentially be replayed year after year and not much has changed in terms of arguments.

1

DeepGamingAI t1_izwolrw wrote

Thanks, that clarifies some things. I have also seen a parameter in the ViT head that simply returns the first token representation instead of averaging across all tokens. I never understood why that made sense, and why only the first token and not some other random token.

This also reminds me of another confusion I have about transformers, would they lose meaning if we gradually compress the embedding size after every mlp in the transformer block?

1

DeepGamingAI t1_izwiy5k wrote

Don't vision transformers do this where instead of gradually compressing the input like a typical convnet, they maintain the high dimensionality throught all the blocks of the deep network, and then simply using a global pooling at the end to compress the "channel" dimension into a compressed representation? I have no idea why that works, but we have seen it does work and the model still learns despite the gradients flowing through this average pooling layer at the end. Would be great if someone can help clarify this for me.

1

DeepGamingAI t1_iuv83lt wrote

>It's about enabling good scholarship and guiding researchers.

You just described the role of a discriminator in a gan

​

>uneducated and unqualified reviewers

op got an invite because they published there before, its on merit not a random review request. besides, the question solely focuses on how reviewing benefits the reviewer, it doesnt seem to cover the whole picture surrounding peer review system

4