Submitted by alkibijad t3_10raouh in MachineLearning
I'm using Huggingface's transformers regularly for experimentations, but I plan to deploy some of the models to iOS.
I have found ml-ane-transformers repo from Apple, which shows how transformers can be rewritten to have much better performance on Apple's devices. There's an example of DistilBERT implemented in that optimized way.
As I plan to deploy transformers to iOS, I started thinking about this. I'm hoping some already have experience about this, so we can discuss:
- Has anyone tried this themselves? Do they actually see the improvements in performance on iOS?
- I'm using Huggingface's transformer models in my experiments. How much work do you think there is to rewrite model in this optimized way?
- It's very difficult to train transformers from scratch (especially if they're big :) ), so I'm fine-tuning on top of pre-trained models on Huggingface. Is it possible to use weights from pretrained Huggingface models with the Apple's reference code? How difficult is it?
[deleted] t1_j6uqyrb wrote
[removed]