dataslacker
dataslacker t1_japjz5y wrote
Reply to comment by [deleted] in [P] InventBot - Invent Original Ideas with Keywords by [deleted]
Sorry but this probably isn’t the right sub for this
dataslacker t1_japi8u8 wrote
This is the least imaginative use of chatGPT I’ve seen yet
dataslacker t1_j926win wrote
Reply to [D] Please stop by [deleted]
Use the down vote button people. I think people just scroll by them
dataslacker t1_j89nmct wrote
Reply to comment by Nameless1995 in [D] The (minor/major?) flaw in the philosophy of OpenAI's ChatGPT purpose by [deleted]
I think this is the correct answer, I very much doubt they are censoring this question of purpose. But the broader question of the types of things they are censoring and wether they are removing bias or adding it is, in my opinion, valid.
dataslacker t1_j7mf8yg wrote
Reply to comment by bitemenow999 in [D] Should I focus on python or C++? by NoSleep19
Interesting! I guess if you’re a well know academic you can get away with that, but the rest of us need to know how to code
dataslacker t1_j7m1xjg wrote
Reply to [D] Should I focus on python or C++? by NoSleep19
Take it from someone who learned C++ first, start with python. You are actually very unlikely to get an interview in C++. The industry standard is Python. Know your algorithms and data structure well enough to do the intermediate level questions on hackerrank and you’ll be in good shape
dataslacker t1_j7m1dho wrote
Reply to comment by bitemenow999 in [D] Should I focus on python or C++? by NoSleep19
I’ve been working in ML for 8 years and I’ve never seen or heard of a scientist being hired without at least one coding interview. Never seen someone just “write down an algorithm” and hand it off to an engineer. I would really like to hear where you saw this because it’s no where near my experience at big tech companies.
dataslacker t1_j7gfa6c wrote
There’s probably some resentment that google and meta could have released something similar over a year ago but chose not to because they didn’t think it would be responsible. Now the company that was founded on being “responsible” released it to the world it a way that hasn’t satisfied a lot of researchers.
dataslacker t1_j4z8zm4 wrote
Reply to comment by JClub in [R] A simple explanation of Reinforcement Learning from Human Feedback (RLHF) by JClub
Yes, your explanations are clear and are also how I understood the paper, but I feel like there's some motivation for the RL training that's missing. Why not "pseudo labeling"? Why is the RL approach better? Also the reward score is non-differentiable because it was designed that way, but they could have designed it to be differentiable. For example instead of decoding the log probs why not train the reward model on them directly? You can still obtain the labels via decoding them doesn't mean that has to be the input to the reward model. There are a number of design choice the authors made that are not motivated in the paper. I haven't read the reference so maybe they are motivated elsewhere in the literature, but RL seems like a strange choice for this problem since there isn't a dynamic environment that the agent is interacting with.
dataslacker t1_j4yraoc wrote
Reply to comment by JClub in [R] A simple explanation of Reinforcement Learning from Human Feedback (RLHF) by JClub
Sorry I think didn’t do a great job asking the question. The reward model, as I understand it, will rank the N generated responses from the LLM. So why not take the top ranked response as ground truth, or a weak label if you’d like and train in a supervised fashion predicting the next token. This would avoid a he RL training which I understand is inefficient and unstable.
dataslacker t1_j4xd5aj wrote
That’s a nice explanation but I’m still unclear as to the motivation for RL. You say the reward isn’t differentiable but since it’s just a label that tells us which of the outputs is best why not simply use that output with supervised training?
dataslacker t1_j2z5slt wrote
Reply to [Discussion] If ML is based on data generated by humans, can it truly outperform humans? by groman434
Depends on how your labels, or more generally the target data distribution, is generated. If it’s generated by human subjectivity yea I would agree. However it’s not hard to think of situations where the labels are the output of a physical process or, in the case of prediction, a future event. In these cases the label is not ambiguous and thus not subject to human interpretation. You also have RL systems that play games against each other and reach super human performance that way. Read about AlphaGo or AlphaStar for example.
dataslacker t1_iyjblnp wrote
I’m going to read this paper in detail but I’m wondering if there’s any insight into why DL methods underperform in TS prediction?
dataslacker t1_ixikye2 wrote
Reply to [D] Schmidhuber: LeCun's "5 best ideas 2012-22” are mostly from my lab, and older by RobbinDeBank
LeCun doesn’t actually credit anyone with those ideas. Likely because they are very broad topics with hundreds of contributors over the years. If Schmidhuber is so arrogant that he’s going to claim them all as his own I don’t blame other scientist for not taking him seriously.
dataslacker t1_iwx2d2c wrote
Reply to [OC] Choose wisely - Warren buffet got paid $704 mln in dividends only while the CEO's total compensation was $24.6 mln by theverybigapple
A bar chart to compare two numbers is unnecessary and therefore not beautiful.
dataslacker t1_ir0qfv8 wrote
Reply to comment by bilby_- in [discussion] Is the future of ML development in low code tools? by bilby_-
In my experience sklearn/pandas is about the right level of “low code” in ML. Anything easier and you’re putting too much under the rug. Plus it’s free.
dataslacker t1_ir0pgi5 wrote
Reply to comment by mystic12321 in [discussion] Is the future of ML development in low code tools? by bilby_-
To your last point, often what people don’t seem to understand is that coding is the easy part of ML development. It’s very rare you’ll find a person who is capable of understanding the details of a model and it’s pitfalls etc, but not able to code in python.
dataslacker t1_japo5b5 wrote
Reply to comment by [deleted] in [P] InventBot - Invent Original Ideas with Keywords by [deleted]
Prompt engineers