Reputation: 143
I trained a model using mxnet framework. The inference time for the model is ~ 9 milliseconds. The model mainly consists of conv layers and uses depthwise separable convolution.
I want to run that model in browser. I converted the model to ONNX format then from
ONNX -> tensorflow -> tensorflowjs.
The inference time for tensorflowjs model ~129 milliseconds.
Any suggestion to improve the performance for the model?
I have also tried ONNXJS but it seems it still has few bugs.
Upvotes: 0
Views: 315
Reputation: 825
Re-architecting would be a possibility since you're dealing with 129ms latency. You would have time to send images to an endpoint (EC2, or SageMaker + API Gateway) running a performant inference server.
Vishaal
Upvotes: 1