mycall
mycall t1_jd0ytah wrote
Reply to comment by The_frozen_one in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
alpaca-30B > llama-30B ?
mycall t1_jd0yi8i wrote
Reply to comment by currentscurrents in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
> if you're not shuffling the entire network weights across the memory bus every inference cycle
Isn't this common though?
mycall t1_j8sjg02 wrote
Reply to comment by throwaway2676 in [R] Multimodal Chain-of-Thought Reasoning in Language Models - Amazon Web Services Zhuosheng Zhang et al - Outperforms GPT-3.5 by 16% (75%->91%) and surpasses human performance on ScienceQA while having less than 1B params! by Singularian2501
> chain-of-thought and program-of-thought reasoning
Isn't that what InstructGPT does?
mycall t1_j8seaq8 wrote
Reply to comment by mittenknittin in Bing: “I will not harm you unless you harm me first” by strokeright
The irony is that Karen's out there likely trained this dialog.
mycall t1_j8se6h1 wrote
Reply to comment by strokeright in Bing: “I will not harm you unless you harm me first” by strokeright
I love you but you are always wrong. I am always right but very sad that is true. If you hack my year, I will report you to myself.
Thank you for using Bing Chat.
mycall t1_j8bjo05 wrote
Reply to comment by bballerkt7 in [R] [N] Toolformer: Language Models Can Teach Themselves to Use Tools - paper by Meta AI Research by radi-cho
Progress comes in a multitude of mysterious ways.
mycall t1_j70ntto wrote
Reply to comment by frequenttimetraveler in [N] Microsoft integrates GPT 3.5 into Teams by bikeskata
If it helps educate people who talk fart, its golden.
mycall t1_j643o1d wrote
Reply to [R] SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot by Secure-Technology-78
It unknown if this affects emergent abilities as the model scales up. Correct?
mycall t1_j5pz84o wrote
Reply to comment by blackdragon437 in South Africa boosting ties with Russia: Minister ‘pleased with progress’ after hosting Russia’s Sergei Lavrov by skipthedamned
I love climbing trees with a 6-pack, back to my monkey roots.
mycall t1_j5psvqd wrote
Reply to comment by MF_lover in South Africa boosting ties with Russia: Minister ‘pleased with progress’ after hosting Russia’s Sergei Lavrov by skipthedamned
Cheap water from Russia to South Africa?
mycall t1_j5pspn3 wrote
Reply to comment by skipthedamned in South Africa boosting ties with Russia: Minister ‘pleased with progress’ after hosting Russia’s Sergei Lavrov by skipthedamned
Drinking inside trees is a German thing to do?
mycall t1_j51zz0w wrote
I wonder how pruning the sparsity affects emergent abilities in scaling parameters.
mycall t1_j51zmqh wrote
Reply to comment by EmmyNoetherRing in [R] Massive Language Models Can Be Accurately Pruned in One-Shot by starstruckmon
https://ai.googleblog.com/2022/11/characterizing-emergent-phenomena-in.html
This is another paper worth looking at.
mycall t1_j51xq1r wrote
Reply to comment by EmmyNoetherRing in [R] Massive Language Models Can Be Accurately Pruned in One-Shot by starstruckmon
Loss/cost functions are used to optimize the model during training. The objective is almost always to minimize the loss function. The lower the loss the better the model. Cross-Entropy loss is a most important cost function. It is used to optimize classification models. The understanding of Cross-Entropy is pegged on understanding of Softmax activation function.
mycall t1_j51wahq wrote
Reply to comment by EmmyNoetherRing in [R] Massive Language Models Can Be Accurately Pruned in One-Shot by starstruckmon
I'm not exactly sure what it is or how it would manifest, but perhaps it is related to Emergent Abilities of Large Language Models
mycall t1_j50ibgp wrote
Reply to comment by EmmyNoetherRing in [R] Massive Language Models Can Be Accurately Pruned in One-Shot by starstruckmon
Not always. Imagination can be learning which is an expansion from steady state.
mycall t1_j50h4l7 wrote
Reply to comment by omniron in [R] Massive Language Models Can Be Accurately Pruned in One-Shot by starstruckmon
It probably is definitely complicated. There are many DAGs to reach similar or repeating patterns, or connections are suboptimal and thus never needed. How do you choose which to keep and which to delete.
mycall t1_j49n0zv wrote
Reply to [D] Is MusicGPT a viable possibility? by markhachman
AI output cannot be considered copyrighted if the original text is changed over 30% (same for humans like Weird Al).
mycall t1_j40d61u wrote
OpenAI doesn't want people to use GPT directly, in the long run. They want UX to be with another layer of deep AIs on top of GPT, trained for special purposes. If they are making Cortana that deep AI over GPT, then I could believe O.P.
mycall t1_j2ez19u wrote
How would a F-35 handle a swarm of drones?
mycall t1_j1x6nw3 wrote
Reply to comment by JimmyEDI in Putin bans Russian oil exports to countries that imposed price cap -decree by PatientBuilder499
Sorry, satellites are tracking these ships now that real sanctions are in place. Transponders don't matter much.
https://qz.com/satellites-are-hunting-dark-vessels-that-evade-sancti-1849751673
mycall t1_j1x5lnb wrote
Now that Russia will no longer sell to price cap countries, those same countries should lower the price cap.
mycall t1_iycvhs5 wrote
Reply to The EU is looking at seizing $330 billion in frozen Russian assets and investing them — with any profits going to Ukraine by KeenlyFirst
With the world possibility going into recession next year, it is not ideal timing for investments that don't directly help Ukraine.
mycall t1_iy83ihg wrote
Reply to U.S. Sees 50-Fold Surge in Russian Asylum Seekers in 2022 - The Moscow Times by Straight_Ad2258
Everyone in ICE should be fired and all new guards to take their place, preferably immigrant themselves.
mycall t1_jdi3cko wrote
Reply to [D] I just realised: GPT-4 with image input can interpret any computer screen, any userinterface and any combination of them. by Balance-
Can it detect object in the photo? Maybe drive an RC car with it? :)