Recent comments in /f/deeplearning
[deleted] t1_jegxv85 wrote
Reply to comment by Trick2206 in Any advanced and updated DL courses? by nuquichoco
[deleted]
ab3rratic t1_jegpq3d wrote
Reply to comment by Ok_Development1633 in Should I continue with this? by Eric-Cardozo
It might be. But then tensorflow/pytorch are out there, too, have documentation and have CPU-only modes.
I say this as someone who coded a number of numeric algorithms (BFGS, Nelder-Mead, etc) from scratch just to understand them, while knowing all along I wasn't competing with real OR libs. On a few occasions, my own implementations were handy for proof-of-concept work when adding 3rd party libs was a hassle.
Ok_Development1633 t1_jegotgs wrote
Reply to comment by ab3rratic in Should I continue with this? by Eric-Cardozo
but could it be useful for a portfolio?
ab3rratic t1_jegoldu wrote
Reply to Should I continue with this? by Eric-Cardozo
If you are still learning, you can continue. But with the knowledge that your library likely won't be competitive with what's already out there.
Trick2206 t1_jegl4b1 wrote
Reply to comment by [deleted] in Any advanced and updated DL courses? by nuquichoco
What are you doing in a deep learning sub?
LiquidDinosaurs69 t1_jeg8yos wrote
Reply to Should I continue with this? by Eric-Cardozo
You should do something else. There are a lot of small C++ nn libraries. To make this one competitive with a real deep learning framework you would need to implement everything with gpu which would be painful. Also, python libraries also have the huge benefit of great data science libraries which make it much more convenient to preprocess data for training networks in python vs cpp.
Additionally there are ways to deploy python models to cpp so there’s not much benefit in training with a cpp library.
x11ry0 OP t1_jefvk0p wrote
Reply to comment by suflaj in State of the art for small dataset next vector prediction models? by x11ry0
Hi, thanks a lot for this very detailed answer. I will have a look at your suggestions. Yes we are looking into the mathematical equations too. We fear that it will be quite un-precise but in other hand I understand very well the issue with the small dataset currently available. We may end up exploring different solutions.
Sa1g t1_jefrkw4 wrote
Reply to comment by adharanda11 in Best way to practice or Learn about practical stuff about deeplearning? by adharanda11
After you have played with scikitlearn, spend some time studying its endpoint (it's like an industry standard), then go on and play with tensoflow/torch
danilo62 OP t1_jefo88t wrote
Reply to comment by --dany-- in [D] Best deal with varying number of inputs each with variable size using and RNN? (for an NLP task) by danilo62
Oh, I hadn't realized that you meant that the embedding would contain information about other features, I get it now. I was referencing an RNN since I thought it was the only option due to the variable input size. Thanks
danilo62 OP t1_jefnror wrote
Reply to comment by -pkomlytyrg in [D] Best deal with varying number of inputs each with variable size using and RNN? (for an NLP task) by danilo62
Yeah I'm gonna try both options (with BERT and the bigger models) but since I'm working with a big dataset I'm not sure I'll be able to use the larger models due to the token and request limits. Thanks for the help
--dany-- t1_jefcm9m wrote
Reply to comment by danilo62 in [D] Best deal with varying number of inputs each with variable size using and RNN? (for an NLP task) by danilo62
The embedding contains all information like sentiment or tf-idf. You just need to train a model to predict trait from post embedding then average over all posts by a person. I didn’t suggest using RNN. Are you sure you were replying my comment?
-pkomlytyrg t1_jef4weq wrote
Reply to comment by danilo62 in [D] Best deal with varying number of inputs each with variable size using and RNN? (for an NLP task) by danilo62
Generally, yes. If you use a model with a long context length (BigBird or OpenAI’s ada02), you’ll likely be fine unless the articles you’re embedding are greater than the token limit. If your using BERT or another, smaller model, you have to chunk/average; that can produce fixed sized vectors but you gotta put the work in haha
[deleted] t1_jeezzdh wrote
Reply to Any advanced and updated DL courses? by nuquichoco
[deleted]
Praise_AI_Overlords t1_jeez5po wrote
Reply to comment by I_will_delete_myself in AI Startup Cerebras releases open source ChatGPT-like alternative models by Time_Key8052
That is very likely. I wonder how this works for multimodality. Weights would probably have to hold more.
danilo62 OP t1_jeeygpe wrote
Reply to comment by --dany-- in [D] Best deal with varying number of inputs each with variable size using and RNN? (for an NLP task) by danilo62
But even then, with the other features (sentiment analysis, tf-idf) how would I feed a vector containing a varying number of tokens and other types of features? I can't see how you would this using an RNN. That is for each post
danilo62 OP t1_jeextek wrote
Reply to comment by -pkomlytyrg in [D] Best deal with varying number of inputs each with variable size using and RNN? (for an NLP task) by danilo62
Oh, so those models are able to produce fixed size embeddings of texts? I wasn't aware of that
I_will_delete_myself t1_jeewl7u wrote
Reply to comment by Praise_AI_Overlords in AI Startup Cerebras releases open source ChatGPT-like alternative models by Time_Key8052
Personally I think the limits with those models is just the amount of information that each weight can hold is limited.
-pkomlytyrg t1_jeeeo36 wrote
Reply to comment by --dany-- in [D] Best deal with varying number of inputs each with variable size using and RNN? (for an NLP task) by danilo62
I would embed the whole post (BigBird or OpenAI embeddings have really long context lengths), and just feed that vector into an RNN. As long as the post is between one and 9000 tokens, the embedding shape will remain the same
IshanDandekar t1_jeecfu0 wrote
Reply to comment by adharanda11 in Best way to practice or Learn about practical stuff about deeplearning? by adharanda11
Tensorflow is used to make neural networks. Start with simpler machine learning algorithms. Look into the Scikit-learn library. Scikit-learn is used for simpler and more commonly used machine learning algorithms. Look into kaggle tutorials for better understanding. Hope this helps!
adharanda11 OP t1_jeec8nu wrote
Reply to comment by IshanDandekar in Best way to practice or Learn about practical stuff about deeplearning? by adharanda11
How will I build projects if I don't know how to use tensorflow?
IshanDandekar t1_jee9uy6 wrote
Reply to comment by adharanda11 in Best way to practice or Learn about practical stuff about deeplearning? by adharanda11
You wouldn't want to go into designing machine learning systems and life cycle right now. Just build projects. Don't get stuck into a cycle of MOOCs
nuquichoco OP t1_jee7a1z wrote
Reply to comment by verboseEqualsTrue in Any advanced and updated DL courses? by nuquichoco
Thanks!
adharanda11 OP t1_jee535h wrote
What are people's thoughts about dive into deep learning book?
adharanda11 OP t1_jee51x0 wrote
Reply to comment by IshanDandekar in Best way to practice or Learn about practical stuff about deeplearning? by adharanda11
I was also thinking about this to gain a basic experience of how projects are made and structured.
_icosahedron t1_jeh3rlg wrote
Reply to Any advanced and updated DL courses? by nuquichoco
I am really enjoying the course by https://dlsyscourse.org. You build a DL library called Needle from scratch throughout the lectures. They also offer notebooks and tests to pass, so it's more than just videos.
It's also by CMU, but the videos are different than the one referenced by /u/verboseEqualsTrue, though I plan on watching that one too.
They cover transformers and other network types, but I haven't gotten that far in the videos yet. I'm still implementing an automatic differentiator.
I have found that Karpathy videos are really good too, though I'm not following them too closely yet.