robobub
robobub t1_jdswai9 wrote
Reply to comment by ngildea in [D] GPT4 and coding problems by enryu42
Indeed, it just made them more efficient so we need less of them and/or less pay for them.
robobub t1_jdst84e wrote
Reply to comment by 0382815 in Why is maths so hard for LLMs? by RadioFreeAmerika
Why? Each of those tokens is O(1) and it is predicting each one incrementally, taking into account the ones it has just generated. So the full answer has taken O(m) where m is the number of tokens.
If it is possible for GPT to do 1+1, it can do a large number of them incrementally. It's not smart enough to do it all the time (you'll have more success if you encourage GPT to have a train of thought reasoning) but it's possible.
robobub t1_jdst1oo wrote
Reply to comment by enryu42 in [D] GPT4 and coding problems by enryu42
> Moreover, I doubt any human programmer will have troubles with the "Beginner" problems, regardless of their specialization.
Have you not heard about how many fail to pass FizzBuzz interview questions?
robobub t1_jdsrlbi wrote
Reply to comment by ArcticWinterZzZ in Why is maths so hard for LLMs? by RadioFreeAmerika
While GPT-4 is autoregressive, it takes into account the tokens it has chosen to generate incrementally. So it is only limited to O(1) if it attempts to answer with the correct answer immediately. It can in theory take O(m) steps, where m is the number of intermediate tokens it predicts.
robobub t1_jdsria4 wrote
Reply to comment by liqui_date_me in [D] GPT4 and coding problems by enryu42
While GPT-4 is autoregressive, it takes into account the tokens it has chosen to generate incrementally. So it is only limited to O(1) if it attempts to answer with the correct answer immediately. It can in theory take O(m) steps, where m is the number of intermediate tokens it predicts.
robobub t1_jdornau wrote
Reply to comment by CptTombstone in [D] "Sparks of Artificial General Intelligence: Early experiments with GPT-4" contained unredacted comments by QQII
How do you square that with these observations on fresh problems from Codeforces (further discussion here
robobub t1_j4n3gcm wrote
Reply to [D] Model for detecting rectangle corners? by hundley10
A couple options off the top of my head
- Add orientation prediction to the bounding box
- Add keypoints for the 4 actual corners as a prediction
- Postprocess boxes with classical techniques, looking for the outermost corners that fit certain properties
- Do everything classically, and deal with the difficulties you have mentioned in your comment.
The first two require annotations of attributes for each box, and will be predicted directly by the model. Though note that you don't have to do this for every label, you can just not train parts of the model when certain attributes are unlabeled.
Both will require some care in modeling, e.g. orientation can have a loss condition at 360 degrees that you'll want to handle, and regressing keypoints can be done well and not well, reference how corners are modeled. And then of course you'll need to postprocess the model's outputs to align/visualize on an image.
robobub t1_isfkv1c wrote
Reply to comment by SeaworthinessBoth391 in [D] Modern MLOps architecture info sources by lifesthateasy
> https://fullstackdeeplearning.com/ ; They have a public discord now, join that
Did you have a link? I'm struggling to find it.
robobub t1_jdwa5wf wrote
Reply to comment by robobub in [D] GPT4 and coding problems by enryu42
Ill add this:
If it is possible for GPT to do 1+1, it can do a large number of them incrementally. It's not smart enough to do it all the time by planning ahead, (you'll have more success if you encourage GPT to have a train of thought reasoning here and here) but it's possible.