VinnyVeritas
VinnyVeritas t1_j57ygl7 wrote
Reply to [D] Did YouTube just add upscaling? by Avelina9X
No they didn't.
VinnyVeritas t1_j3rrzvr wrote
Reply to comment by soupstock123 in Building a 4x 3090 machine learning machine. Would love some feedback on my build. by soupstock123
Actually I've been sort of looking at ML computers (kind of of browsing and dreaming one day I would have one, but it's always going to be out of my means and needs anyway). Anyway, they can put two PSUs in a box, obviously it's made by companies, so the total cost is twice or 3 times the cost of the parts alone (e.g. building yourself would be 2-3x cheaper) but it could inspire you for picking your parts.https://bizon-tech.com/amd-ryzen-threadripper-up-to-64-cores-workstation-pc
https://shop.lambdalabs.com/gpu-workstations/vector/customize
VinnyVeritas t1_j3nh3g4 wrote
Reply to comment by soupstock123 in Building a 4x 3090 machine learning machine. Would love some feedback on my build. by soupstock123
I suppose one PSU will take care of motherboard + CPU + some GPUs and the other one will take care of remaining GPUs.
So if you get 4x 3090, that's 350W x4 = 1400W just for GPUs, +300 watts for CPU, +powering the rest of the components, drives, etc... So let's say we round that up to 2000W, then add at least 10% margin, that's 2200W total.
So maybe 1600W PSU for mobo and some GPUs, and another 1000W or more for the remaining GPUs. Note, if you go with 3090TI, it's more like 450-500W per card, so you have to do the maths.
Or if you want to go future proof, just put two 1600W PSUs, and then you can just swap your 3090 with 4090 in the future and not worry about upgrading PSUs.
VinnyVeritas t1_j3ng2u9 wrote
Reply to comment by qiltb in Building a 4x 3090 machine learning machine. Would love some feedback on my build. by soupstock123
Do you have some numbers or a link because all benchmarks I've seen point to the contrary? I'm happy to update my opinion if things have changed and there's data to support it.
VinnyVeritas t1_j3l0gqt wrote
Reply to Building a 4x 3090 machine learning machine. Would love some feedback on my build. by soupstock123
I don't know if that's going to work well to have 16 PCIe lane, everyone here I've seen making 4 GPUs machines uses the CPUs that have 48 or 16 PCIe lanes.
Also you'll need a lot of watts to power that monster, not to mention you need a 10-20% margin if you don't want fry the PSU.
VinnyVeritas t1_j3l04w8 wrote
Reply to comment by hjups22 in Building a 4x 3090 machine learning machine. Would love some feedback on my build. by soupstock123
Each time someone asks this question, someone repeats this misinformed answer.
This is incorrect, NVLink doesn't make much difference.
VinnyVeritas t1_j0xc7l5 wrote
Reply to comment by TheMrZZ0 in Biggest 3090 deep learning rigs? 4x ? 8x? 64x? by Outrageous_Room_3167
Thanks that makes sense, I thought they were a startup in the business of building computers, I was completely confused!!!
VinnyVeritas t1_j0w0gvh wrote
I'm not following: you're doing start-up on infrastructure build and you have to ask for advice on reddit to scale past 1 machine? That gives a terrible image of your startup. To the average person like me it sounds like you don't know what you're doing.
VinnyVeritas t1_izk4aoa wrote
Wow, that's a nice extension/replacement. Very cool work!
VinnyVeritas t1_iya34at wrote
Correlation is not causation.
VinnyVeritas t1_iy2nq1w wrote
Reply to Best GPU for deep learning by somebodyenjoy
Just get two 4090 and power limit them to 350 watts using "nvidia-smi", there's almost no performance loss and you don't need NVLink anyway to do multi-gpu training.
VinnyVeritas t1_ixzu8ny wrote
Reply to Deep Learning for Computer Vision: Workstation or some service like AWS? by Character-Ad9862
Those $10,000 won't last long on AWS. There's also LambdaLabs, their cloud prices are a lot more affordable. They also make dedicated servers for machine learning.
VinnyVeritas t1_ixs2kfa wrote
Reply to comment by jazzzzzzzzzzzzzzzy in Is Linux still vastly preferred for deep learning over Windows? by moekou
Maybe it was a pain years ago, I don't really know but nowadays you just click install nVidia drivers in the software management and it works. There's nothing painful or difficult about it.
VinnyVeritas t1_ixj2wd6 wrote
Reply to [D] Schmidhuber: LeCun's "5 best ideas 2012-22” are mostly from my lab, and older by RobbinDeBank
If I was Schmidhuber, instead of whining every time someone "stole my idea that didn't really work" and made it work, I'd just revisit my old ideas and make them work myself.
It's like inventing the airplane, there's a big difference between the general idea and actually making one that flies.
VinnyVeritas t1_iwreqw0 wrote
Reply to [R] RWKV-4 7B release: an attention-free RNN language model matching GPT-J performance (14B training in progress) by bo_peng
Really cool stuff!
VinnyVeritas t1_iwfnzp9 wrote
Reply to [P] 🔥 CleanRL has reached v1.0.0; Reworked documentation, JAX support, and more! by vwxyzjn
This is amazing, the code is so clear I feel I can finally understand RL!
Bookmarked.
VinnyVeritas t1_iwelmj4 wrote
Looks like a lot of sprinting...
VinnyVeritas t1_iw9ajwe wrote
Reply to comment by master3243 in [R] ZerO Initialization: Initializing Neural Networks with only Zeros and Ones by hardmaru
The performance is not better: the results are the same within the margin of error for standard (not super-deep networks). Here I copied from their table:
Cifar10
ZerO Init 5.13 ± 0.08
Kaiming Init 5.15 ± 0.13
Imagenet
ZerO Init 23.43 ± 0.04
Kaiming Init 23.46 ± 0.07
VinnyVeritas t1_iw1s40n wrote
Reply to comment by master3243 in [R] ZerO Initialization: Initializing Neural Networks with only Zeros and Ones by hardmaru
Like what? Training ultra-deep neural networks without batchnorm? But in their experiments the accuracy gets worse with deeper networks, what's the point of going deeper to get worse results?
VinnyVeritas t1_iw0w76i wrote
Seems useless, why not simply fix the seed of the random generator for reproducibility?
VinnyVeritas t1_ivqdjyf wrote
Reply to comment by gahaalt in Pytorch Symbolic: an equivalent of Keras Functional API [Project] by gahaalt
Oh I didn't realize you saw it as a simplification because all it does is remove the input size but then you need to create input placeholders. I thought it was meant to reimplement Keras on PyTorch.
VinnyVeritas t1_ivnv1dt wrote
I am not understanding the point of rewriting Keras on top of PyTorch? Is it just for your own fun?
Why not use Keras directly? Keras has better deployment capabilities as a bonus.
VinnyVeritas t1_jcdelqr wrote
Reply to [D] What do people think about OpenAI not releasing its research but benefiting from others’ research? Should google meta enforce its patents against them? by [deleted]
Ultimately AI will become a big boys club, where big corporate will hold all the cards.
OpenAI just made the first leap towards that dystopian near future.