Torch-TensorRT Backend for torch.compile #1690
Replies: 1 comment
-
This won't help much if it continues to be seemingly impossible to install pytorch-tensorrt. I get the same version 0.0.0 installation problem I see opened over and over again here with no resolution. If torch.compile(backend="tensorrt") doesn't yet work should torch_tensorrt.compile() work? And if it does how do I install the frustratingly difficult to install thing? I keep reading more and more so called resolutions and none actually resolve anything. |
Beta Was this translation helpful? Give feedback.
-
Torch-TensorRT Backend for torch.compile
TL;DR
PyTorch 2.0's inference story is centralized around torchdynamo and torch.compile. It provides an interface and compiler stack to optimize models and therefore a natural entry point to integrate the FX frontend.
Goal(s)
Provide a natural interface to all of Torch-TensorRT's features through
torch.compile
Usecases
Proposed APIs / UX
Users would interact with torch.compile as an API for torch-tensorrt.
Example Workflow
Limitations
There shouldnt be any difference between
torch.compile
andtorch-tensorrt.compile
Internal Implementation
Design
The main component of the implementation is the definition of the backend which invokes intermediary steps such as AOTAutorgrad.
Extensions Required to Core API implementations
There should be few changes to the core API
Data Structures
There might be ways to make it easier for users to group settings like using the compile spec object.
Details specific for TorchScript Support
N/A
Details specific for FX support
See above
Implementation Phases
Prototype - M
MVP
(<TARGET RELEASE VERSION>)
- SBeta Was this translation helpful? Give feedback.
All reactions