I have been using TensorRT and TensorFlow-TRT to accelerate the inference of my DL algorithms.
Then I have heard of:
Both seem to accelerate DL. But I am having a hard time to understand them. Can anyone explain them in simple terms?
I have been using TensorRT and TensorFlow-TRT to accelerate the inference of my DL algorithms.
Then I have heard of:
Both seem to accelerate DL. But I am having a hard time to understand them. Can anyone explain them in simple terms?
Trax
is a deep learning framework created by Google and extensively used by the Google Brain team. It comes as an alternative toTensorFlow
andPyTorch
when it comes to implementing off-the-shelf state of the art deep learning models, for example Transformers, Bert etc. , in principle with respect to the Natural Language Processing field.Trax
is built uponTensorFlow
andJAX
.JAX
is an enhanced and optimised version of Numpy. The important distinction aboutJAX
andNumPy
is that the former using a library called XLA (advanced linear algebra) which allows to run yourNumPy
code onGPU
andTPU
rather than onCPU
like it happens in the plainNumPy
, thus speeding up computation.