Soubhi M. Hadri
Soubhi M. Hadri

Reputation: 143

Speed-up tensorflowjs model

I trained a model using mxnet framework. The inference time for the model is ~ 9 milliseconds. The model mainly consists of conv layers and uses depthwise separable convolution.

I want to run that model in browser. I converted the model to ONNX format then from

ONNX -> tensorflow -> tensorflowjs.

The inference time for tensorflowjs model ~129 milliseconds.

Any suggestion to improve the performance for the model?

I have also tried ONNXJS but it seems it still has few bugs.

Upvotes: 0

Views: 315

Answers (1)

Vishaal
Vishaal

Reputation: 825

Re-architecting would be a possibility since you're dealing with 129ms latency. You would have time to send images to an endpoint (EC2, or SageMaker + API Gateway) running a performant inference server.

Vishaal

Upvotes: 1

Related Questions