C0hentheBarbarian

C0hentheBarbarian t1_j9o5068 wrote

I work in NLP. Work mainly consists of fine tuning NLP models. With the rise of LLMs I'm seeing a lot of my work becoming Prompt engineering. I'm happy to pick up the new skill but I'd like to know what avenues I have to upskill beyond being a prompt engineer without a PhD. Feels like all the learning I did on model architectures etc is going to waste. There are still a few projects that need me to fine tune a model for text classification etc but as LLMs get better I suspect I need better skills to go beyond becoming a prompt engineer. For anyone else in NLP who doesn't have a PhD and doesn't have any experience building model architectures/training from scratch etc, how are all of you trying to up skill in these times? EDIT: Worded the question to ask only people who don't have a PhD, I would actually like to know everyone's perspective on this.

1

C0hentheBarbarian t1_iyh4kex wrote

Results like this make me seriously question if I'll have a job in the future as an ML person. I understand the nature of the job will change etc but I can see myself becoming an overqualified prompt engineer.

31

C0hentheBarbarian t1_is96iqz wrote

Highly recommend this post by Jay Alammar. He has one of the best tutorials on how transformers work too (IMO) and this one is up there. I have worked with CV very sporadically recently but his post along with some of the links he has on there explained things to me pretty well. The only math background I can recommend off the top of my head is the probability calculation for lower/upper bounds - you can look up how VAEs work there or the post I linked has resources to understand the same.

2