Ulfgardleo
Ulfgardleo t1_j1vd1op wrote
Reply to comment by KonArtist01 in [P] Can you distinguish AI-generated content from real art or literature? I made a little test! by Dicitur
there are definitely signs. paper texture is often wrong. hands are often wrong. With all my guesses of "old master" i was never really sure, but with the AI guesses i often were pretty confident.
Ulfgardleo t1_j1vcqri wrote
Reply to comment by respeckKnuckles in [P] Can you distinguish AI-generated content from real art or literature? I made a little test! by Dicitur
-
you are asking humans to solve this task untrained, which is not the same as the human ability to distinguish the two.
-
you are then also making it harder by phrasing the task in a way that makes it difficult for the human brain to solve it.
Ulfgardleo t1_j1afdic wrote
Reply to comment by [deleted] in How do fusion scientists expect to produce enough Tritium to sustain D-T fusion (see text)? by DanTheTerrible
other projects do not "completely dismiss" this issue, but they are researching other aspects of fusion. For example: that Wendelstein-7x does not investigate breeding is not because they don't see it as interesting, but because they want to show that stellarator confiment works and that we have finally understood plasma physics.
Ulfgardleo t1_iza5481 wrote
the hessian is always better information than the natural gradient because it includes actual information of the curvature of the function while the NG only includes curvature of the model. So any second order TR approach with NG information will approach the hessian.
//edit: I am assuming actual trust region methods, like TR-Newton, and not some RL-ML approximation schemes.
Ulfgardleo t1_iz9fjio wrote
Reply to comment by bohreffect in [R] The Forward-Forward Algorithm: Some Preliminary Investigations [Geoffrey Hinton] by shitboots
Funny that stuff always comes back. We used to differentiate SVM solutions wrt kernel parameters like that back in the day.
Ulfgardleo t1_iz437fi wrote
Reply to comment by chaosmosis in [R] The Forward-Forward Algorithm: Some Preliminary Investigations [Geoffrey Hinton] by shitboots
I have a story to tell about the one time where i got invited as external evaluator for a MSc thesis. I agreed, later opened it and then realized it was a comparison of 10 animal migration algorithms.
This thesis sat on my desk for WEEKS because i did not know how to grade it. How do you grade pseudo science?!? Like, it is not the fault of the students to fall prey to this topic, but I also can't condone them not figuring out that it IS pseudoscience.
Ulfgardleo t1_iz3q7pd wrote
Reply to comment by _der_erlkonig_ in [R] The Forward-Forward Algorithm: Some Preliminary Investigations [Geoffrey Hinton] by shitboots
I did not. I did it for Hinton.
A heuristic can be useful without proof, especially in tasks that are very difficult to solve. However, you have to supply strong theoretic arguments why they should work. A biological analog is not enough, especially if it is one that we do not understand, either.
Otherwise you end up like the other category of nature inspired optimization heuristics that pretend to optimize by mimicking the hunting patterns of the Harris hawk. And I wished I made this up just now.
Ulfgardleo t1_iz2ampb wrote
Reply to [R] The Forward-Forward Algorithm: Some Preliminary Investigations [Geoffrey Hinton] by shitboots
I will start believing in Hinton's algorithms once they proof that it is consistent with some vector field with fixed points that are meaningful optima of some objective function.
Ulfgardleo t1_iylr164 wrote
Reply to comment by notdelet in [R] Statistical vs Deep Learning forecasting methods by fedegarzar
The "and" in the post you replied to was a logical "and". The best evaluation procedure does not help if you use poor, underperforming baselines.
Ulfgardleo t1_ixhlloe wrote
Reply to comment by RobbinDeBank in [D] Schmidhuber: LeCun's "5 best ideas 2012-22” are mostly from my lab, and older by RobbinDeBank
but he does not claim that. What he does claim is that developments in his lab predate those ideas. It might be that those ideas were rediscovered independently by others, but as in so many things, who is first matters. And from a scientific point of view, not doing literature review for ones own ideas is bad science, and especially so if a lab strategically avoids citing some other lab.
It shouldn't be so difficult to write "idea X [1] rediscovered by [more prominent 2] has led to" in ones work.
Ulfgardleo t1_iwzpfez wrote
Reply to comment by datasciencepro in [D] David Ha/@hardmaru of Stability AI is liking all of Elon Musk's tweets by datasciencepro
then unfollow him.
Ulfgardleo t1_ivxodj0 wrote
Reply to comment by DreamyPen in [Discussion] Can we train with multiple sources of data, some very reliable, others less so? by DreamyPen
-
okay you completely confuse everyone in the ML community when you call inputs "labels". lets keep with inputs/outputs.
-
This is good, because it allows you to estimate some crude measure for the quality of the physics model.
So, label noise is a broad field. I am mostly knowledgeable in the classification setting, where label noise has different effects. Moreover, you are not in the standard noisy label setting, because the noise is not independent of the label, so just using weights will be difficult. Similarly, if you have more than one output to predict, a single weight is difficult to compute.
The standard way to derive all of these methods is by noting that the MSE can be derived as the log-probability of the normal distribution p(y|f) where y is the ground truth and f is the mean, and variance is some fixed value. For the mse, the value of the variance does not matter, as long as it remains fixed, but with fairly little effort you can show that as soon as you give samples individual variances, this amounts to weighting the MSE.
So, the cheapest approach would be to give outcomes from the different sources a different variance and if you have more than one output, you will also have more than one variance. How do you guess the parameters? well, make them learnable parameters and train them together with your model parameters.
Of course you can make it arbitrarily complicated. Since your cheap labels come from a physics simulation, errors are likely correlated so you can learn a full covariance matrix. And from there you can make it as complex as you like by making the error distribution more complex, but you will likely not have enough data to do so.
Ulfgardleo t1_ivvjvsd wrote
Reply to comment by DreamyPen in [Discussion] Can we train with multiple sources of data, some very reliable, others less so? by DreamyPen
- you said "unreliable outputs". did you mean inputs? if you truely meant outputs (i.e., the material properties that you want to predict from some so far undefined inputs) then this is what in ml is called "label".
- Okay, i have the same issue here. typically ground-truth would be what we called label, but i can see that you would distinguish between simulated/measured ground-truth data.
- "model" here is the physics based model, not the ML-model, right?
- I don't see it answered. I ask it explicitely: is there any experimental measurmeent for which you also have the physics-model output?
- You lost me here.
Ulfgardleo t1_ivvfwzf wrote
Reply to comment by Far-Butterscotch-436 in [Discussion] Can we train with multiple sources of data, some very reliable, others less so? by DreamyPen
it is not so easy. if we talk about noise in the input patterns and not in the labels then the noise inputs can be catastrophic to model performance. in that case the model needs to know what data source the input is from.
Ulfgardleo t1_ivvfosf wrote
Reply to [Discussion] Can we train with multiple sources of data, some very reliable, others less so? by DreamyPen
Okay, there are a few questions:
- what is unreliable: the inputs, or the labels? Is your problem even supervised?
- What do you want to learn?
- Is it possible to quantify reliability for each source? Is it just higher variance or also bias?
- Does there exist cases for which you have reliable and unreliable data?
- What is the data you finally predict on? the reliable or unreliable data?
Ulfgardleo t1_ivtuenl wrote
It is such a rare opportunity to get a better labeled dataset. If i was still working it, i would use this dataset to evaluate noisy label techniques.
Ulfgardleo t1_ivlw2dk wrote
Reply to comment by Monarch252001 in [D] Academia: The highest funded plagiarist is also an AI ethicist by [deleted]
Because it is suboptimal to ask people to Google when casually throwing around one of the biggest accusations in science.
Ulfgardleo t1_iurwhof wrote
Reply to comment by ReginaldIII in [R] Is there any work being done on reduction of training weight vector size but not reducing computational overhead (eg pruning)? by Moose_a_Lini
hi, you are aware i am not the OP?
Ulfgardleo t1_iuqhizr wrote
Reply to comment by bernhard-lehner in [R] Is there any work being done on reduction of training weight vector size but not reducing computational overhead (eg pruning)? by Moose_a_Lini
you might have misunderstood the message. OP asked if there is work on reducing the number of weights without compromising network strength as it is not compute that is the bottleneck but transmitting the object.
Ulfgardleo t1_itzmz4u wrote
Reply to comment by SleekEagle in [D] Poisson Flow Generative Models - new physics inspired generative model by SleekEagle
my comment was just a remark saying that "diffusion model" by itself is not informative, because there have been several approaches that already brought a lot of speed-up. The standard diffusion model is not state-of-the-art anymore.
Ulfgardleo t1_ityn2zg wrote
i am not sure i buy the 10-20x faster claim as stated in the form of this post (but i did not have time to read the link, yet). This is because there are diffusion models that claim to be 10-40x faster than diffusion models.
Ulfgardleo t1_iswx9zs wrote
Reply to comment by DorianGre in [D] How frustrating are the ML interviews these days!!! TOP 3% interview joke by Mogady
(a+b-c)/(d+e)
4 function calls in numpy.
Ulfgardleo t1_isec031 wrote
Reply to How are quantized vectors vectors? by [deleted]
technically that holds true for all real number representations of vectors we have. your computer computes stuff in "pretend" vector spaces and it just happens to approximately work out. Sometimes you have to care about the numerics, but most of the time you are fine.
​
//edit Having said that, there are ways to make it mathematically work. A vector space is typically defined over a field, e.g., the real or complex numbers, and in this case the scalars are limited to that field. The same construction over a ring (like the integers) is called a module. So a quantized vector (for example the integers) limited to scalars being integers as well DOES work out. The only thing you have to deal with is that divison in a ring might not exist.
Ulfgardleo t1_irv9w00 wrote
Reply to comment by gradientrun in [D] What are your thoughts about weak supervision? by ratatouille_artist
quite easy to proof.
take a multi-class classification problem. Now, pick one class and assign it label 0, assign all other classes the same coarse label 1 and try to find the maximum margin classifier. This problem is equivalent to finding a convex polytope that separates class 0 from class 1 with maximum margin. This is an NP-hard problem. Logistic regression is not much better, but more difficult to proof.
This is already NP-complete when the coarse label encompasses two classes: https://proceedings.neurips.cc/paper/2018/file/22b1f2e0983160db6f7bb9f62f4dbb39-Paper.pdf
Ulfgardleo t1_j1vjc6q wrote
Reply to comment by respeckKnuckles in [P] Can you distinguish AI-generated content from real art or literature? I made a little test! by Dicitur
I don't think this is one of those cases. The question we want to answer is whether texts are good enough that humans will not pick up on it. Making the task as hard as possible for humans is not indicative of real world performance once people get presented these texts more regularly.