ソースを参照

fix incorrect link in triton readme

Swetha Mandava 5 年 前
コミット
3d6e3d612c
1 ファイル変更2 行追加3 行削除
  1. 2 3
      TensorFlow/LanguageModeling/BERT/triton/README.md

+ 2 - 3
TensorFlow/LanguageModeling/BERT/triton/README.md

@@ -1,4 +1,3 @@
-
 # Deploying the BERT TensorFlow model using Triton Inference Server
 
 This folder contains instructions for deployment and exemplary client application to run inference on
@@ -183,7 +182,7 @@ For more information about `perf_client`, refer to the [official documentation](
 
 ### Latency vs Throughput for TensorRT Engine
 
-Performance numbers for BERT Large, sequence length=384 are obtained from [experiments]([https://github.com/NVIDIA/TensorRT/tree/release/7.1/demo/BERT#inference-performance-nvidia-a100-40gb](https://github.com/NVIDIA/TensorRT/tree/release/7.1/demo/BERT#inference-performance-nvidia-a100-40gb)) on NVIDIA A100 with 1x A100 40G GPUs. Throughput is measured in samples/second, and latency in milliseconds.
+Performance numbers for BERT Large, sequence length=384 are obtained from [experiments](https://github.com/NVIDIA/TensorRT/tree/release/7.1/demo/BERT#inference-performance-nvidia-a100-40gb) on NVIDIA A100 with 1x A100 40G GPUs. Throughput is measured in samples/second, and latency in milliseconds.
 
 ![](../data/images/bert_trt_throughput_vs_latency.png?raw=true)
 
@@ -232,4 +231,4 @@ April 2020
 TRTIS -> TRITON
 
 October 2019
-Initial release
+Initial release