Submitted by avialex t3_xzq6bs in MachineLearning
I was doing some looking into QML (specifically quantum neural networks) this past week, checking out a few of the python frameworks like Qiskit and Pennylane, and I've come up against this fundamental problem with QNN's right now. Because of the no-cloning theorem, you cannot save state. Because you cannot save state, you cannot do backprop. And because you can't do backprop, you have to run the model twice for each parameter, varying it to find the derivative (whether you're using finite diff or parameter shift). So classical NN's are O(1) per data, you do one feedforward and one backprop per input, where with QNN's it's O(n)!!
So, that seems completely insane. Leaving aside the technical problems and assuming we could just build a n-scale quantum system with no decoherence noise, a "small" classical model like Resnet18 has 11,000,000 trainable parameters. That's 22 million forward passes just to do one training step for the QNN variation. My thought is, there must be some advantage I'm missing. Like, that the representational ability of a coherent quantum system must be enormous compared to classical. But all the examples I've seen of equivalent-size QNN's to NN's show pretty much equivalent performance.
What's up then? Is there an advantage at scale that we haven't realized yet? Or are they hoping to find some technique of finding the gradients that takes much less time? Am I just missing something?
pmirallesr t1_irnmvpg wrote
You're thinking about implementing classic ml (backprop) in qc. Proponents of quantum ML look for alternative ways of "machine learning", either not calculating gradients, or trying to exploit the properties of quantum mechanics to "learn better". If it all sounds very fuzzy it's because it is