fnbr
fnbr t1_jd6j7gh wrote
Right now, the tech isn't there to train on a single GPU. You're gonna end up training a language model for ~1 month to do so. It is slightly more efficient, though.
Lots of people looking at running locally. In addition to everything that people have said, there's a bunch of companies that will be releasing models that can just barely fit on an A100 soon that I've heard rumours about from employees.
fnbr t1_j7q9v1m wrote
Reply to [D] List of RL Papers by C_l3b
Sutton and Barto is obligatory if you want to learn RL, imo. Even as an experienced researcher, I read it every year or so. It's very approachable.
fnbr t1_j3uzyom wrote
Become a manager, you’ll have tons of meetings with people.
Alternatively, get a social hobby. You don’t need social connections to come from work.
fnbr t1_j1bc8i3 wrote
The main problem with tabular Q-learning (I'm assuming that by classical, you mean tabular) is that for most environments that are interesting, the state space is massive, so we can't actually store all states in memory.
In particular for lunar lander, you have a continuous observation space, so you need to apply some sort of discretization; at that point, you might as well just use tile coding or some sort of other function approximator.
fnbr t1_iv9iag6 wrote
Why force the network to do this instead of just sorting the predictions afterward?
fnbr t1_iuj8h11 wrote
Reply to [D] When the GPU is NOT the bottleneck...? by alexnasla
Have you profiled your code? That would be the first thing I would do.
What sort of utilization of the GPU are you getting?
It's likely you're bottlenecked by feeding data in- for supervised learning, that's often the case.
I'm happy to offer suggestions for feeding data in if you're using Tensorflow/JAX.
fnbr t1_iuhn5pu wrote
Reply to comment by SignificantPen1325 in Tax Return Question About Investment Losses by SignificantPen1325
If you do this, make sure you don’t buy the same security within 30 days or it’s a “wash-loss” and you can’t dedicate the loss from your income.
fnbr t1_iqortww wrote
What sort of loads are you planning on running on your laptop? If you’re running LLMs the answer’s gonna be totally different than if you’re trying to train RL agents in grid worlds.
fnbr t1_jd6jsb6 wrote
Reply to [D] What's the Time and Space Complexity of Transformer Models Inference? by Smooth-Earth-9897
The best analysis on this is from this blog:
https://kipp.ly/blog/transformer-inference-arithmetic/