DeepLearningExamples/FasterTransformer
2020-03-02 14:10:33 +01:00
..
v1 [FasterTransformer] Adding v2 2020-03-02 14:10:33 +01:00
v2 [FasterTransformer] Adding v2 2020-03-02 14:10:33 +01:00
README.md [FasterTransformer] Adding v2 2020-03-02 14:10:33 +01:00

FasterTransformer

This repository provides a script and recipe to run the highly optimized transformer for inference, and it is tested and maintained by NVIDIA.

Table Of Contents

Model overview

FasterTransformer V1

FasterTransformer V1 provides a highly optimized BERT equivalent Transformer layer for inference, including C++ API, TensorFlow op and TensorRT plugin. The experiments show that FasterTransformer V1 can provide 1.3 ~ 2 times speedup on NVIDIA Tesla T4 and NVIDIA Tesla V100 for inference.

FasterTransformer V2

FastTransformer V2 adds a highly optimized OpenNMT-tf based decoder and decoding for inference in FasterTransformer V1, including C++ API and TensorFlow op. The experiments show that FasterTransformer V2 can provide 1.5 ~ 11 times speedup on NVIDIA Telsa T4 and NVIDIA Tesla V 100 for inference.

Architecture matrix

The following matrix shows the Architecture Differences between the model.

Architecure Encoder Decoder
FasterTransformer V1 Yes No
FasterTransformer V2 Yes Yes

Release notes

FasterTransformer V1 will be deprecated on July 2020.

Changelog

February 2020

  • Release the FasterTransformer 2.0
  • Provide a highly optimized OpenNMT-tf based decoder and decoding, including C++ API and TensorFlow OP.
  • Refine the sample codes of encoder.
  • Add dynamic batch size feature into encoder op.

July 2019

  • Release the FasterTransformer 1.0
  • Provide a highly optimized bert equivalent transformer layer, including C++ API, TensorFlow OP and TensorRT plugin.

Known issues

There are no known issues with this model.