blimpyway
blimpyway t1_j744ina wrote
Reply to comment by Feeling_Card_4162 in [R] Topologically evolving new self-modifying multi-task learning algorithms by Feeling_Card_4162
If you use an evolutionary algorithm like NEAT what is the selection criteria?
blimpyway t1_j742oes wrote
Reply to [D] Why do LLMs like InstructGPT and LLM use RL to instead of supervised learning to learn from the user-ranked examples? by alpha-meta
I guess the point of the reward model is to approximate human feedback and instead of hiring humans to actually rank (e.g.) 1billion chats needed to update the LLM, train a reward model with 1% of them then use it to simulate human evaluators 99% of the times.
blimpyway t1_j73zt6s wrote
Reply to [R] Topologically evolving new self-modifying multi-task learning algorithms by Feeling_Card_4162
So what they (it?) evolve for?
blimpyway t1_j6tmu2t wrote
Reply to comment by Soft-Material3294 in [R] SETI finds eight potential alien signals with ML by logTom
Alliens are naturally occurring too
blimpyway t1_j51wv3h wrote
Retrieval should work also on entire interaction history with a particular user. Not only tracking beyond token window but having available all "interesting stuff" from users perspective.
blimpyway t1_j4ulemc wrote
Reply to [P] RWKV 14B Language Model & ChatRWKV : pure RNN (attention-free), scalable and parallelizable like Transformers by bo_peng
Prior to this, have you experimenting with smaller (== more manageable) variants of this model or previous variants were attempted directly at this scale?
blimpyway t1_j4pqj1b wrote
Reply to [D] Is it possible to update random forest parameters with new data instead of retraining on all data? by monkeysingmonkeynew
Search for "online random forests" there quite a few papers and articles on the subject. I assume they aren't simple usage of normal RF-s, it wouldn't worth publishing a paper on the subject.
blimpyway t1_j4pndcs wrote
Reply to comment by blackkettle in [R] The Predictive Forward-Forward Algorithm by radi-cho
One application I can think of is learning on edge. There is an industry fashion to embed AI inference capabilities in the newer ARM chips. The so called NPUs. Which are simplified GPUs optimized only for inference (forward passes). Such an algorithm would enable them to learn using only forward passes, hence without requiring backpropagation.
Another possibility I think is ability to train one layer at a time, which diminishes GPU memory requirements.
And probably more important it opens the gates for all kind of not yet seen network architectures, topologies and training methods that do not require fully differentiable pathways.
edit: regarding the brain inspired part.. well you can dismiss it as AI's reversed cargo cult - if it imitates some properties of the brain it should act like the brain, but I would be cautious to attribute Hinton this kind of thinking. Brains are very different from ANNs and trying to emulate their properties could provide insights on how they work.
blimpyway t1_j4pi713 wrote
Reply to comment by chaosmosis in [D] The Illustrated Stable Diffusion (Video) by jayalammar
The order of the words/tokens is normally encoded via positional embeddings that are added each to their respective token embedding. See e.g. https://machinelearningmastery.com/a-gentle-introduction-to-positional-encoding-in-transformer-models-part-1/
blimpyway t1_j4hbgtt wrote
Reply to comment by derpderp3200 in [R] from a human motion sequence, SUMMON synthesizes physically plausible and semantically reasonable objects by t0ns0fph0t0ns
> David Lewandowski videos It looks like they search for some furniture
blimpyway t1_j36r8vz wrote
Reply to comment by Ill_Spread_6434 in The Persistent Problem of Consciousness: an astronaut's epiphany by simsquatched
Or maybe neither. Consciousness could simply be the inwards reference frame.
Something akin to outwards reference frame yet having an opposite direction.
blimpyway t1_j1aqukr wrote
Reply to Google tells employees more of them will be at risk for low performance ratings next year by Sorin61
Couldn't that be a natural consequence of having more employees next year?
blimpyway t1_j1aqaza wrote
I don't know, what were your results?
blimpyway t1_izidx41 wrote
Reply to comment by nova9001 in UK, Italy and Japan team up to develop a new fighter jet that uses artificial intelligence by marketrent
Basically it's a sky Roomba
blimpyway t1_iz08hbs wrote
Reply to comment by frstyle34 in Study in fruit flies found that the most promising anti-ageing drug rapamycin prolongs the lifespan, and the development of age-related pathological changes in the gut, of female fruit flies, but not in males. Concluding that the sex is a crucial factor in the effectiveness of anti-ageing drugs by giuliomagnifico
Changing it.
blimpyway t1_ixr5ro9 wrote
Reply to Nighttime artificial outdoor lighting was associated with impaired glucose control and 28% higher increased diabetes risk, a cross-sectional study on ~100k people in China showed by giuliomagnifico
Seriously? Couldn't be the case the outdoor night lights only shine the road towards the candy shop?
blimpyway t1_iwz31ut wrote
Reply to [P] Any object detection library by PegasusInvasion
And how would an image without any object look like?
blimpyway t1_ivjly2a wrote
Reply to comment by f10101 in [D] At what tasks are models better than humans given the same amount of data? by billjames1685
This indeed could be one case. However a couple hundred attempts is not the limit - a kid would get it in less than a couple dozen trials or she will get bored.
However I found that some models can do it even faster. Like under 5 failures or less on 50% trials, including only 2 failures in 5% of trials.
blimpyway t1_ivjl2sn wrote
Reply to comment by Pawngrubber in [D] At what tasks are models better than humans given the same amount of data? by billjames1685
> One easy example: if you heavily leverage tree search algorithms and have a tiny neural net eval (much smaller than stockfish nnue) it would still surpass humans even with only hundreds of games.
Any reference on that?
blimpyway t1_ivjkpvi wrote
Reply to [D] At what tasks are models better than humans given the same amount of data? by billjames1685
Yes for all living organisms their learning competencies are geared towards sample efficiency. Whoever spends a bit too much on figuring out how to solve a problem either starves to death and more likely is eaten long before that.
blimpyway t1_ivizpp3 wrote
Reply to comment by evanthebouncy in [D] At what tasks are models better than humans given the same amount of data? by billjames1685
Is any ML model able to (learn how to) solve that? With pen and paper I can, and cannot credit 500M yr of genetic heritage for it.
blimpyway t1_iviwvnt wrote
Reply to comment by ramblinginternetnerd in [D] At what tasks are models better than humans given the same amount of data? by billjames1685
Now let's figure out how to store such a pretrained model in ~700MBytes of genetic code, without disturbing the other info about how all non-brainy enzymes, organs and tissues, etc.. should be built and assembled together.
blimpyway t1_ivivcwr wrote
Reply to comment by IntelArtiGen in [D] At what tasks are models better than humans given the same amount of data? by billjames1685
Tesla collected 780M miles of driving till 2016
A human learning to drive for 16h/day at an average speed of 30mph for 18years would have a data set of ~3M miles.
So we can say humans are at least 1000 times more sample efficient than whatever Tesla and any other autonomous driving companies are doing.
blimpyway t1_jebtlmt wrote
Reply to State of the art for small dataset next vector prediction models? by x11ry0
Assemble a dataset and raise a challenge on Kaggle?