Submitted by mjrossman t3_107i7yv in Futurology
unholyravenger t1_j3rwzvs wrote
Reply to comment by Helpful_Opinion2023 in A Singular Trajectory: the Signs of AGI by mjrossman
For a conceptual understanding start with 3b1b . One of the best explanations of the underlying concepts I've seen. This is really the foundation of everything.
Next, there are 2 main concepts to understand, and that is how each layer of a NN works and the overall architecture. A quick list of layers to get your head around: Linear also called Multilayer Perceptron (MLP), CNN Convolutional Neural Network, then you have a family of layers that handles sequences like sentences. These are RNN, LSTM, and Transformers. But all of these are built on the same concepts as the 3b1b videos. If you're more of a math person this is a great way to conceptualize what each of these layers is doing.
Next different architectures. Start will simple classifiers, which you should already have a good understanding of. Then check out how GAN's work and how you can use two networks to train each other. Then maybe you can go to the state of the art with Diffusion networks. I think this is a bit easier to understand than how each layer works.
All the while playing around in python, and prepackaged ML stuff to apply your knowledge to something concrete. Make a simple classifier, download, and fine-tune a diffusion network on some dataset. Coursera has some really good classes, particularly by Andrew NG who is one of the biggest ML educators out there.
​
Bonus Resources:
ML Streat Talk: Podcast talking to people in the industry. Lots of deep concepts here.
2min Papers The hype channel. Learn about all the new stuff coming out.
Yannic Kilcher: Go deep into different papers written on ML and how they work at a very deep level.
​
Good luck it's a lot, but no one knows everything and you need surprisingly little to get started.
Viewing a single comment thread. View all comments