Closed
Description
Bug Description
We optimized one model based on TRTorch v0.3.0 successfully, but failed to deploy in online-serving, which maybe core-dump or hang at MemoryD2H. By locking torch.jit inference,we could work-around this issue currently, and we guess trtorch runtime is not thread-safe.
I collected other issue or commits related this problem:
- trtorch community: ❓ [Question] Is the module compiled by TRTorch thread safe? #181
- TFTRT had fixed this problem by this patch: tensorflow/tensorflow@e51fa30
From these issue or commits, I get nvinfer1::IExecuteContext is not thread-safe.
So how to fix this bug in trtorch?