BrisklyBrusque
BrisklyBrusque t1_j7g8v7a wrote
Reply to comment by maxip89 in [N] "I got access to Google LaMDA, the Chatbot that was so realistic that one Google engineer thought it was conscious. First impressions" by That_Violinist_18
I don’t think it was marketing. The individual who claimed the AI was sentient was dismissed for his actions.
BrisklyBrusque t1_j43dsux wrote
Reply to [D] Are there any papers on optimization-based approaches which combine learned parameter initializations with learned optimisers? by Decadz
You might enjoy “Well-Tuned Simple Nets Excel on Tabular Data”
https://arxiv.org/abs/2106.11189
Authors wrote a computer routine that leverages BOHB (Bayesian optimization and Hyberband) to search an enormous search space of possible neural network architectures. The authors allowed the routine to select different regularization techniques, including many ensemble techniques like dropout, snapshot ensembles, and others that render the choice of parameter initializations less critical. However, authors used the same optimizer (AdamW) in all experiments.
Not exactly what you are looking for but hopefully interesting.
BrisklyBrusque t1_j2wwoym wrote
Reply to [D] Is there any research into using neural networks to discover classical algorithms? by currentscurrents
Some of the earliest perceptron research involved emulating logic gates with single and multilayer perceptrons. A single layer perceptron famously could not learn XOR gates but a multilayer perceptron could.
BrisklyBrusque t1_j0i9ldc wrote
Reply to comment by chaosmosis in [R] Are there open research problems in random forests? by SpookyTardigrade
Thanks for the suggestion.
BrisklyBrusque t1_j0hx440 wrote
Yes, lots. For example, in 2019 a paper introduced a new split rule for categorical variables that reduces computational complexity.
https://peerj.com/articles/6339/
A lot of researchers are also exploring adjacent tree ensembles such as extremely randomized trees (2006) and Bayesian additive regression trees (2008). The former is very similar to random forests. There is a strong possibility other tree ensembles have yet to be discovered!
If you’re a fan of computer science / optimized code, there is a great deal of research concerning making tree models faster. The ranger library in R was introduced as an improvement on the randomForest package. There is also interest in making random forests scale up to millions of variables, to deal with genetics data.
Hummingbird is a Microsoft project that seeks to refactor common machine learning methods using tensor algebra, so those methods can take advantage of GPUs. I don’t know if they got around to random forests yet.
Random forests raise a lot of questions about the relationship between ensemble diversity and ensemble accuracy, about which there are many mysteries.
BrisklyBrusque t1_iyj6bja wrote
Reply to comment by CyberPun-K in [R] Statistical vs Deep Learning forecasting methods by fedegarzar
13 days to tune multiple deep neural networks is not at all unrealistic depending on the number of gpus.
BrisklyBrusque t1_iy8wfoa wrote
Reply to comment by olmec-akeru in [D] What method is state of the art dimensionality reduction by olmec-akeru
I freely admit I haven’t looked into the math. But my understanding was the embeddings are a learned representation. They are not arbitrary; instead they aim to put categories close to one another on a continuous scale only in those situations where it is justified.
BrisklyBrusque t1_iy3slot wrote
Well sometimes dimension reduction is used to maintain the most important aspects of the data in as few vectors as possible, particularly when we want to visualize high-dimensional data or escape the curse of dimensionality. Other times dimension reduction is more of a regularization technique. Think of self-organizing maps, RBMs, autoencoders, and other neural nets that learn a representation of the data, which can then be passed to another neural net as the new training sample.
So dimension reduction is itself a technique with many distinct applications.
BrisklyBrusque t1_iy3s0ha wrote
Reply to comment by olmec-akeru in [D] What method is state of the art dimensionality reduction by olmec-akeru
Cool links. I’ll add “entity embeddings” into the mix. Entity embeddings reimagine a categorical variable as a continuous-valued vector and allow us to skip one-hot encoding.
BrisklyBrusque t1_iv6otss wrote
Reply to comment by BeatLeJuce in [D] What are the major general advances in techniques? by windoze
I recall that experimenters disagreed on why batchnorm worked in the first place? has the consensus settled?
BrisklyBrusque t1_iv6ogqg wrote
2007-2010: Deep learning begins to win computer vision competitions. In my eyes, this is what put deep learning on the map for a lot of people, and kicked off the renaissance we see today.
2016ish: categorical embeddings/entity embeddings. For tabular data with categorical variables, categorical embeddings are faster and more accurate than one-hot-encoding, and preserve the natural relationships between factors by mapping them to a low dimensional space
BrisklyBrusque t1_iv6negg wrote
Reply to comment by cautioushedonist in [D] What are the major general advances in techniques? by windoze
Is this different from the premise that neural networks are universal function approximators?
BrisklyBrusque t1_is7g6gk wrote
Reply to comment by MOSFETBJT in [D] Are GAN(s) still relevant as a research topic? or is there any idea regarding research on generative modeling? by aozorahime
SVMs prevailed against neural networks in a big image classification contest in 2006. Then they fell out of favor, with other learning algorithms like
•Adaboost
•C4.5
•Decision stumps
•Multivariate adaptive regression splines
•Flexible discriminant analysis
•Arcing
•Wagging
Not sure which of these will come back, but it’s funny how often ideas are rediscovered (like neural networks themselves, which were branded as multilayer perceptrons initially)
BrisklyBrusque t1_is7fdk7 wrote
Reply to [D] Are GAN(s) still relevant as a research topic? or is there any idea regarding research on generative modeling? by aozorahime
Anything is relevant as a research topic if you’re passionate about it. Just be prepared to stand up for yourself when you go against the grain. The harshest critics will label your research useless if they are convinced that better methods are available.
BrisklyBrusque t1_j7kwj3i wrote
Reply to comment by kaidon in [OC] Movie production companies with the highest total worldwide box office revenue by giteam
When Disney acquired Fox a few years ago, it became pretty clear Disney would be getting about 1/3 of all the revenue that goes to the box office. pretty insane.