|
|
5 лет назад | |
|---|---|---|
| .. | ||
| v1 | 6 лет назад | |
| v2 | 5 лет назад | |
| v2.1 | 5 лет назад | |
| LICENSE | 6 лет назад | |
| README.md | 5 лет назад | |
This repository provides a script and recipe to run the highly optimized transformer for inference, and it is tested and maintained by NVIDIA.
FasterTransformer V1 provides a highly optimized BERT equivalent Transformer layer for inference, including C++ API, TensorFlow op and TensorRT plugin. The experiments show that FasterTransformer V1 can provide 1.3 ~ 2 times speedup on NVIDIA Tesla T4 and NVIDIA Tesla V100 for inference.
FastTransformer V2 adds a highly optimized OpenNMT-tf based decoder and decoding for inference in FasterTransformer V1, including C++ API and TensorFlow op. The experiments show that FasterTransformer V2 can provide 1.5 ~ 11 times speedup on NVIDIA Telsa T4 and NVIDIA Tesla V 100 for inference.
FasterTransformer V2.1 optimizes some kernels of encoder and decoder, adding the support of PyTorch, the support of remove the padding of encoder and the support of sampling algorithm in decoding.
The following matrix shows the Architecture Differences between the model.
| Architecure | Encoder | Decoder | Decoding with beam search | Decoding with sampling |
|---|---|---|---|---|
| FasterTransformer V1 | Yes | No | No | No |
| FasterTransformer V2 | Yes | Yes | Yes | No |
| FasterTransformer V2.1 | Yes | Yes | Yes | Yes |
FasterTransformer V1 will be deprecated on July 2020.
FasterTransformer V2 will be deprecated on Dec 2020.
June 2020
May 2020
translate_sample.py.April 2020
decoding_opennmt.h to decoding_beamsearch.hdecoding_sampling.hbert_transformer_op.h, bert_transformer_op.cu.cc into bert_transformer_op.ccdecoder.h, decoder.cu.cc into decoder.ccdecoding_beamsearch.h, decoding_beamsearch.cu.cc into decoding_beamsearch.ccbleu_score.py into utils. Note that the BLEU score requires python3.March 2020
translate_sample.py to demonstrate how to translate a sentence by restoring the pretrained model of OpenNMT-tf.February 2020
July 2019
There are no known issues with this model.