I'm trying to follow the instructions in this guide:
https://docs.aws.amazon.com/dlami/latest/devguide/tutorial-inferentia-tf-neuron.html and compile a savedModel with Neuron.
When the model is large (>2GB) I'm failing on ValueError: Message tensorflow.GraphDef exceeds maximum protobuf size of 2GB error.
I understand this is a hard limit on the model size from Tensorflow, but is there any workaround from Neuron end? Is there any solution for running a large model on AWS Inferentia?
Thanks.