소스 검색

[BERT/PyT] specify GPU for triton (#666)

Sharath T S 5 년 전
부모
커밋
8588e9834c
1개의 변경된 파일1개의 추가작업 그리고 1개의 파일을 삭제
  1. 1 1
      PyTorch/LanguageModeling/BERT/triton/README.md

+ 1 - 1
PyTorch/LanguageModeling/BERT/triton/README.md

@@ -102,7 +102,7 @@ To make the machine wait until the server is initialized, and the model is ready
 
 ## Performance
 
-The numbers below are averages, measured on Triton, with [static batching](https://docs.nvidia.com/deeplearning/sdk/tensorrt-inference-server-guide/docs/model_configuration.html#scheduling-and-batching). 
+The numbers below are averages, measured on Triton on V100 32G GPU, with [static batching](https://docs.nvidia.com/deeplearning/sdk/tensorrt-inference-server-guide/docs/model_configuration.html#scheduling-and-batching). 
 
 | Format | GPUs | Batch size | Sequence length | Throughput - FP32(sequences/sec) | Throughput - mixed precision(sequences/sec) | Throughput speedup (mixed precision/FP32)  |
 |--------|------|------------|-----------------|----------------------------------|---------------------------------------------|--------------------------------------------|