But most of the papers focus on Collaborative Filtering -> MF. My assumption is that this approach has are potential? Please tell whether this is a right assumption.
They deploy the models with well established pipelines. Usually (from my experience), scientists communicate closely with engineers (ML, Data Engineers or DevOps). They "put" the model into the environment where it can operate. It can be either the web service where u simply wrap the model, or container deployments (see Docker). They don't retrain the models every-time the app launches as they simply save the models parameters in that environment or CI/CD platform. I have a little experience in deploying and maintaining the production models, but I have no clue about the details tho.
Not really, it is about approximating (or better to say mapping) of real-world values that brings the limits. So that the model shrinks - computations and other model operations are being executed at lower bit-width(s).
Awekonti OP t1_j0z1ads wrote
Reply to comment by __lawless in [D] Deep Learning based Recommendation Systems by Awekonti
Yes, I will also take a look at GNN and Diffusion models based. Thank you)