zbyte64
zbyte64 t1_jdmvaak wrote
Reply to comment by Blacky372 in [D] Do we really need 100B+ parameters in a large language model? by Vegetable-Skill-9700
Sounds like we're realizing that a model is only as good as the experts that wrote the training data.
zbyte64 t1_jcdzvhh wrote
Reply to comment by Philpax in [N] PyTorch 2.0: Our next generation release that is faster, more Pythonic and Dynamic as ever by [deleted]
That's why all my ML is done in OvjectiveC /s. Production looks different for different use cases.
zbyte64 t1_j8zfbi0 wrote
Write a bot to handle all HR complaints and train it on the latest managerial materials. Then as a bonus the bot will look at all the conversations and propose metrics for increased efficiency and harmony at the work place.
zbyte64 t1_j74y5o9 wrote
Reply to [R] Multimodal Chain-of-Thought Reasoning in Language Models - Amazon Web Services Zhuosheng Zhang et al - Outperforms GPT-3.5 by 16% (75%->91%) and surpasses human performance on ScienceQA while having less than 1B params! by Singularian2501
What kind of hardware do I need to train this?
zbyte64 t1_j28z09i wrote
Reply to comment by Tip_Odde in Your wastewater can be used as renewable energy by EnergyTransitionNews
Idk, magic bullets sound expensive. I'm sure parent would do the economically sensible thing and allow themselves to be eaten by the werewolf.
zbyte64 t1_iw5e6hv wrote
"Under the same data" - I guess that rules out applying this to the plethora of models popping up under stable diffusion
zbyte64 t1_jeeps93 wrote
Reply to comment by tripple13 in [D][N] LAION Launches Petition to Establish an International Publicly Funded Supercomputing Facility for Open Source Large-scale AI Research and its Safety by stringShuffle
Let me guess, it's not about any of those things and they're really a lobbying mechanism to engage in protectionism for the industry.