MisterManuscript
MisterManuscript t1_jdhrqhl wrote
Reply to Cuda out of memory error by Rishh3112
What GPU are you using? How much vRAM does it have?
MisterManuscript t1_jdbuih4 wrote
Reply to comment by tdgros in [R] Introducing SIFT: A New Family of Sparse Iso-FLOP Transformations to Improve the Accuracy of Computer Vision and Language Models by CS-fan-101
I stand corrected regarding the patent. The naming conflict, on the other hand, is here to stay.
MisterManuscript t1_jdawa9m wrote
Reply to [R] Introducing SIFT: A New Family of Sparse Iso-FLOP Transformations to Improve the Accuracy of Computer Vision and Language Models by CS-fan-101
Feels like the authors are trying to piggyback on the pre-existing fame of Scale-Invariant Feature Transform. Out of all other names that could have been chosen, why try to override an existing name?
Addendum: if you're lucky, Google just might cut you some slack. If not, then expect their lawyers to come at you with a cease-and-desist.
Addendeum 2: from a deleted reply from one of the authors/person from Cerebras asking why Google might come after them with a cease-and-desist: SIFT's patent is owned by Google. They may consider trademark violation, or something similar.
MisterManuscript t1_jcxo0zv wrote
You don't need the 40 series, they're designed with providing ML solutions to games. You're paying extra just to have an in-built optical-flow accelerator that you're not gonna use for model training.
The optical flow accelerator is meant for computing dense optical flow fields as part of many inputs to the DLSS feature that most new games use.
You're better off with the 30 series or lesser.
MisterManuscript t1_jcry6cj wrote
Reply to comment by MysteryInc152 in [R] ChatGLM-6B - an open source 6.2 billion parameter Eng/Chinese bilingual LLM trained on 1T tokens, supplemented by supervised fine-tuning, feedback bootstrap, and RLHF. Runs on consumer grade GPUs by MysteryInc152
That's not what bootstrapping is, it is a resampling technique used to create multiple datasets of the same size from the original dataset using random sampling with replacement. It is done to get the estimate of the standard deviation of a desired variable.
Here's the link to the ISLR textbook. The bootstrap chapter will verify what it is.
MisterManuscript t1_jcrwvc8 wrote
Reply to comment by Temporary-Warning-34 in [R] ChatGLM-6B - an open source 6.2 billion parameter Eng/Chinese bilingual LLM trained on 1T tokens, supplemented by supervised fine-tuning, feedback bootstrap, and RLHF. Runs on consumer grade GPUs by MysteryInc152
I tried googling it, it's is a nonexistent terminology in the realm of statistics. I know what bootstrapping is, but not this version of it.
It's better to ask the GitHub authors about this to make sure they're not just spitting out pseudostatistical terminology.
Addendum: another guy did query the authors regarding this terminology in the issues tab, they did not respond.
MisterManuscript t1_j58tvbk wrote
Reply to [D] Did YouTube just add upscaling? by Avelina9X
Could be from simple upscaling techniques (e.g. bilinear interpolation) combined with antialiasing. CNNs don't seem feasible given the compute power needed for the number of users using YouTube.
MisterManuscript t1_izwkkoj wrote
If this is an ML model, what is the input? I can't find the format of the input in the article here. Is it 2D images? Text prompt?
MisterManuscript t1_ivxatob wrote
Reply to [Discussion] Suggestions on how to annotate X-ray images with only radiology reports available by AJ521432
You'll need to consult a doctor specialised in respiratory medicine. Pretty sure no one in ML can teach you how annotate medical images unless they are trained in medicine.
MisterManuscript t1_iv460q1 wrote
That extra money spent on the RTX 4090 is gonna go to waste since you're not using the the built-in optical-flow accelerators for training; they're designed and optimized specifically, and only to efficiently compute optical-flow fields for games. Better off sticking to RTX 3090.
MisterManuscript t1_isrnlam wrote
Reply to comment by Overall-Importance54 in [D] What is the deal with breast cancer scans? by Overall-Importance54
You can check out SIIM-ISIC Melanoma classification competition.
MisterManuscript t1_isosuxa wrote
Reply to comment by Overall-Importance54 in [D] What is the deal with breast cancer scans? by Overall-Importance54
That's a different use case. I work in healthcare (computer vision solutions) and I have yet to see a use-case for text-to-video generation.
So far most of the use-cases are either: classification or segmentation, mostly for diagnosing diseases. There's also biomedical informatics, but I'm not familiar enough with medical tabular data to comment on it.
MisterManuscript t1_isobcjb wrote
This isn't unique to your case. A lot of applications of machine learning in healthcare suffer from one drawback: a lot of patient data is sensitive and is very hard to get your hands on. You don't just query them to your liking; there is a lot of bureaucracy in place before you get your hands on even a small sample of required data.
MisterManuscript t1_jdhtan0 wrote
Reply to comment by Rishh3112 in Cuda out of memory error by Rishh3112
You probably have a memory leak somewhere in youe training loop. Either that or your model or batch size is way too big and occupies a lot of vRAM.
Addendum: There's a difference between RAM and vRAM (your GPU's RAM), I hope the 14GB you're talking about is vRAM and not the RAM of your AWS vm.