Submitted by PassingTumbleweed t3_10qzlhw in MachineLearning
Every day, there seems to be new evidence of the generalization capabilities of LLMs.
What does this mean for the future role of deep learning experts in academia and business?
It seems like there's a significant chance that skills such as PyTorch and Jax will be displaced by prompt construction and off-the-shelf model APIs, with only a few large institutions working on the DNN itself.
Curious to hear others' thoughts on this.
uchi__mata t1_j6srpnd wrote
I don't see prompt construction obviating the need for coding skills, even as the prompts improve I still think you're going to want knowledgeable humans to review the scripts before using them in critical apps, but I do think tools like GPT will rapidly speed up prototyping and eliminate boilerplate dev for most engineers.
That said, model APIs strike me as a much more likely disruptor of workaday software dev because as they prove themselves out it'll just make financial sense for firms to have fewer people creating bespoke models vs pulling stuff off the shelf and modifying it as needed. In this world data science largely becomes an orchestration task with ML ops/data engineering + understanding of business need and available data being translated into ML pipeline creation to solve problems. People working directly on model creation from scratch would mostly be academics and highly skilled CS/stats/math PhDs working at a handful of large tech companies and model API firms. This seems like the most probable future to me as almost every innovation in tech goes this route eventually.
Basically, if a task doesn't require deep understanding of business needs, it's subject to commoditization.