zzzthelastuser
zzzthelastuser t1_izx8k9l wrote
Reply to comment by master3243 in [D] - Has Open AI said what ChatGPT's architecture is? What technique is it using to "remember" previous prompts? by 029187
> I've seen discussion on whether they use some kind of summarization to be able to fit more context into the same
They could unironically use ChatGPT for this task.
zzzthelastuser t1_iwpltkw wrote
Reply to comment by ReasonablyBadass in [R] Will we run out of data? An analysis of the limits of scaling datasets in Machine Learning - Epochai Pablo Villalobos et al - Trend of ever-growing ML models might slow down if data efficiency is not drastically improved! by Singularian2501
Sure, but in theory my little Hello World network had also more data available on the internet.
zzzthelastuser t1_iwpi7r5 wrote
Reply to comment by ReasonablyBadass in [R] Will we run out of data? An analysis of the limits of scaling datasets in Machine Learning - Epochai Pablo Villalobos et al - Trend of ever-growing ML models might slow down if data efficiency is not drastically improved! by Singularian2501
You could argue GPT-3 was trained on a subset of the available training data, no?
Not completing the first pass-through means the remaining data could be considered as not part of the training data.
zzzthelastuser t1_iw7nx3a wrote
Reply to [D] ML/AI role as a disabled person by badhandml
I know many researchers who barely program/know how to program aside from scripting and tweaking stuff they found on github. So I would say you will be fine!
But admittedly I think it wont be as easy to enter these jobs without prior job experience.
zzzthelastuser t1_iui1g14 wrote
Reply to comment by starstruckmon in [P] Explain Paper - A Better Way to Read Academic Papers by xutw21
It might, that's why I'm asking.
Also "just use GPT3" is bit underselling the whole interface around it.
zzzthelastuser t1_iuhlntw wrote
Will this remain free? Who is paying for servers etc? Any chance to make this open-source for hosting it locally?
zzzthelastuser t1_itzqczw wrote
Reply to comment by jshkk in [D] What platform environment would you use for young Python ML learners? by jshkk
I've NEVER experienced that colab wasn't available for me and I use it all the time.
zzzthelastuser t1_isnvnv9 wrote
Reply to comment by Overall-Importance54 in [D] What is the deal with breast cancer scans? by Overall-Importance54
Similar with self-driving cars, they may work with (made up number) 99% accuracy, but that 1% is still too risky.
Regardless of what the AI says, I would still ask a doctor to see my scan considering a false-negative could cost me my life and a false-positive would probably mean a doctor would double check it anyway.
The bottleneck would still be the person who looks at each scan personally.
That being said, I think there is huge potential in early risk prediction using ML long before a real human could even spot cancer tissue.
zzzthelastuser t1_is0oiz5 wrote
Reply to comment by ggerganov in [P] Pure C/C++ port of OpenAI's Whisper by ggerganov
I THINK it is possible to keep the file cached. So that if a user returns to the site the model doesn't need to be re-downloaded again.
Alternatively a user could download the model file manually and your website asks the user to drag and drop their model file to launch the service?
zzzthelastuser t1_j7ulu8h wrote
Reply to [P] Get 2x Faster Transcriptions with OpenAI Whisper Large on Kernl by pommedeterresautee
> CUDA graphs require us to capture a graph per input tensor shape, there is a non-negligible warmup time. We measure around 10mn on 2 different machines / GPUs (down from 50mn in our previous Kernl version). One user reported with the new version a bit more than 20mn of warmup time. We are aware of obvious ways to decrease it significantly.
Dumb question, but what's mn? millineconds?