SleekEagle

SleekEagle OP t1_jaszawj wrote

It looks like, rather than conditioning on text they condition on the fMRI, but it's unclear to me exactly how they map between the two and why this would even work without finetuning. TBH I haven't had time to read the paper so I don't know the details, but figured I'd drop the paper in case anyone was interested!

7

SleekEagle t1_j9vl7r3 wrote

I don't think anyone believes it will be LLMs that undergo an intelligence explosion, but they could certainly be a piece of the puzzle. Look at how much progress has been made in just the past 10 years alone - imo it's not unreasonable to think that the alignment problem will be a serious concern in the next 30 years or so.

In the short term, though, I agree that people doing bad things with AI is much more likely than an intelligence explosion.

Whatever anyone's opinion, I think the fact that the opinions of very smart and knowledgeable people run the gamut is a testament to the fact that we need to dedicate serious resources into ethical AI beyond the disclaimers at the end of every paper that models may contain biases.

2

SleekEagle t1_j9tttxr wrote

Until the tools start exhibiting behavior that you didn't predict and in ways that you have no control over. Not taking an opinion on which side is "right", just saying that this is a false equivalence with respect to the arguments that are being made.

​

EDIT: Typo

6

SleekEagle t1_j8ix4fz wrote

Authors publish papers on research, experiments, findings, etc. They do not always release the code for the models they are studying.

The lucidrains' repos implement the models, creating an open-source implementation for the research

The next step would then be to train the model, which requires a lot more than just the code (most notably, money). I assume you're referring to these trained weights when you say "the needed AI model". Training would require a huge amount of time and money for a team, never mind a single person, to train even one of these models let alone a whole portfolio of them

For this reason, it's not very reasonable to expect lucidrains or any other person to train these models - the open-source implementations are a great contribution on their own!

9

SleekEagle t1_iy4k5m4 wrote

It's been a while since I looked at tsne and umap but the assumption for PCA is that the data lives near an affine subspace and for VAE that the data is well modeled by the distribution whose parameters you are finding. My thoughts but I'm sure there's other considerations that I'd love to hear other people chime in with!

6

SleekEagle OP t1_iued3mr wrote

To generate data, you need to know the probability distribution of a dataset. This is in general unknown. The method called "normalizing flows" starts with a simple distribution that we do know exactly, and learns how to turn the simple distribution into the data distribution through a series of transformations. If we know these transformations, then we can generate data from the data distribution by sampling from the simple distribution and passing it through the transformations.

Normalizing flows are a general approach to generative AI - how to actually learn the transformations and what they look like depends on the particular method. With PFGMs, the authors find that the laws of physics define these transformations. If we start with a simple distribution, we can transform it into the data distribution by imagining the data points are electrons and moving them according to the electric field they generate.

2

SleekEagle OP t1_iu5h5tt wrote

I'm not sure how the curse of dimensionality would affect PFGMs relative to Diffusion Models, but at the very least PFGMs could be dropped in as the base model in Imagen while diffusion models are kept for the super resolution chain! More info on that here or more info on Imagen here (or how to build your own Imagen here ;) ).

2

SleekEagle OP t1_iu569xx wrote

I don't think the paper explicitly says anything about this, but I would expect them to be similar. If anything I would imagine they would require less memory, but not more. That having been said, if you're thinking of e.g. DALL-E 2 or Stable Diffusion, those models also have other parts that PFGMs don't (like text encoding networks), so it is completely fair that they are larger!

4