diff --git a/.github/scripts/generate-tensorrt-test-matrix.py b/.github/scripts/generate-tensorrt-test-matrix.py index e5a9b9fae4..7ff90c7ffd 100644 --- a/.github/scripts/generate-tensorrt-test-matrix.py +++ b/.github/scripts/generate-tensorrt-test-matrix.py @@ -5,13 +5,15 @@ import json import sys +import requests # type: ignore[import-untyped] + # please update the cuda version you want to test with the future tensorRT version here # channel: nightly if the future tensorRT version test workflow is triggered from the main branch or your personal branch # channel: test if the future tensorRT version test workflow is triggered from the release branch(release/2.5 etc....) CUDA_VERSIONS_DICT = { - "nightly": ["cu124"], - "test": ["cu121", "cu124"], - "release": ["cu121", "cu124"], + "nightly": ["cu126"], + "test": ["cu124", "cu126"], + "release": ["cu124", "cu126"], } # please update the python version you want to test with the future tensorRT version here @@ -29,39 +31,78 @@ "10.4.0": { "urls": "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.4.0/zip/TensorRT-10.4.0.26.Windows.win10.cuda-12.6.zip", "strip_prefix": "TensorRT-10.4.0.26", - "sha256": "3a7de83778b9e9f812fd8901e07e0d7d6fc54ce633fcff2e340f994df2c6356c", }, "10.5.0": { "urls": "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.5.0/zip/TensorRT-10.5.0.18.Windows.win10.cuda-12.6.zip", "strip_prefix": "TensorRT-10.5.0.18", - "sha256": "e6436f4164db4e44d727354dccf7d93755efb70d6fbfd6fa95bdfeb2e7331b24", }, "10.6.0": { "urls": "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.6.0/zip/TensorRT-10.6.0.26.Windows.win10.cuda-12.6.zip", "strip_prefix": "TensorRT-10.6.0.26", - "sha256": "6c6d92c108a1b3368423e8f69f08d31269830f1e4c9da43b37ba34a176797254", + }, + "10.7.0": { + "urls": "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.7.0/zip/TensorRT-10.7.0.23.Windows.win10.cuda-12.6.zip", + "strip_prefix": "TensorRT-10.7.0.23", }, }, "linux": { "10.4.0": { "urls": "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.4.0/tars/TensorRT-10.4.0.26.Linux.x86_64-gnu.cuda-12.6.tar.gz", "strip_prefix": "TensorRT-10.4.0.26", - "sha256": "cb0273ecb3ba4db8993a408eedd354712301a6c7f20704c52cdf9f78aa97bbdb", }, "10.5.0": { "urls": "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.5.0/tars/TensorRT-10.5.0.18.Linux.x86_64-gnu.cuda-12.6.tar.gz", "strip_prefix": "TensorRT-10.5.0.18", - "sha256": "f404d379d639552a3e026cd5267213bd6df18a4eb899d6e47815bbdb34854958", }, "10.6.0": { "urls": "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.6.0/tars/TensorRT-10.6.0.26.Linux.x86_64-gnu.cuda-12.6.tar.gz", "strip_prefix": "TensorRT-10.6.0.26", - "sha256": "33d3c2f3f4c84dc7991a4337a6fde9ed33f5c8e5c4f03ac2eb6b994a382b03a0", + }, + "10.7.0": { + "urls": "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.7.0/tars/TensorRT-10.7.0.23.Linux.x86_64-gnu.cuda-12.6.tar.gz", + "strip_prefix": "TensorRT-10.7.0.23", }, }, } +def check_new_tensorrt_version( + major: int = 10, patch_from: int = 0 +) -> tuple[bool, str, str, str, str]: + def check_file_availability(url: str) -> bool: + try: + response = requests.head(url, allow_redirects=True) + if response.status_code == 200: + content_type = response.headers.get("Content-Type", "") + content_disposition = response.headers.get("Content-Disposition", "") + if "application" in content_type or "attachment" in content_disposition: + return True + return False + except requests.RequestException: + return False + + trt_linux_release_url = "" + trt_win_release_url = "" + + # calculate the next minor version + minor = int(list(TENSORRT_VERSIONS_DICT["linux"].keys())[-1].split(".")[1]) + 1 + trt_version = f"{major}.{minor}.0" + for patch in range(patch_from, 50): + for cuda_minor in range(4, 11): + trt_linux_release_url_candidate = f"https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/{trt_version}/tars/TensorRT-{trt_version}.{patch}.Linux.x86_64-gnu.cuda-12.{cuda_minor}.tar.gz" + if check_file_availability(trt_linux_release_url_candidate): + trt_linux_release_url = trt_linux_release_url_candidate + trt_win_release_url = f"https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/{trt_version}/zip/TensorRT-{trt_version}.{patch}.Windows.win10.cuda-12.{cuda_minor}.zip" + return ( + True, + trt_version, + str(patch), + trt_linux_release_url, + trt_win_release_url, + ) + return False, "", "", "", "" + + def main(args: list[str]) -> None: parser = argparse.ArgumentParser() parser.add_argument( @@ -99,6 +140,25 @@ def main(args: list[str]) -> None: f"{includes[0].validation_runner} is not the supported arch, currently only support windows and linux" ) + ( + new_trt_available, + trt_version, + trt_patch, + trt_linux_release_url, + trt_win_release_url, + ) = check_new_tensorrt_version(major=10, patch_from=0) + if new_trt_available: + TENSORRT_VERSIONS_DICT["linux"][trt_version] = {} + TENSORRT_VERSIONS_DICT["linux"][trt_version]["urls"] = trt_linux_release_url + TENSORRT_VERSIONS_DICT["linux"][trt_version][ + "strip_prefix" + ] = f"TensorRT-{trt_version}.{trt_patch}" + TENSORRT_VERSIONS_DICT["windows"][trt_version] = {} + TENSORRT_VERSIONS_DICT["windows"][trt_version]["urls"] = trt_win_release_url + TENSORRT_VERSIONS_DICT["windows"][trt_version][ + "strip_prefix" + ] = f"TensorRT-{trt_version}.{trt_patch}" + cuda_versions = CUDA_VERSIONS_DICT[channel] python_versions = PYTHON_VERSIONS_DICT[channel] tensorrt_versions = TENSORRT_VERSIONS_DICT[arch] diff --git a/.github/scripts/generate_binary_build_matrix.py b/.github/scripts/generate_binary_build_matrix.py index 26bb447b4f..0db7073889 100644 --- a/.github/scripts/generate_binary_build_matrix.py +++ b/.github/scripts/generate_binary_build_matrix.py @@ -64,7 +64,7 @@ XPU = "xpu" -CURRENT_NIGHTLY_VERSION = "2.6.0" +CURRENT_NIGHTLY_VERSION = "2.7.0" CURRENT_CANDIDATE_VERSION = "2.5.1" CURRENT_STABLE_VERSION = "2.5.1" CURRENT_VERSION = CURRENT_STABLE_VERSION diff --git a/.github/workflows/assigner.yml b/.github/workflows/assigner.yml index 2b65e554b1..b1056c50b3 100644 --- a/.github/workflows/assigner.yml +++ b/.github/workflows/assigner.yml @@ -22,7 +22,7 @@ jobs: runs-on: ubuntu-latest steps: - name: Checkout - uses: actions/checkout@v3 + uses: actions/checkout@v4 - name: Assign uses: ./.github/actions/assigner diff --git a/.github/workflows/build-tensorrt-linux.yml b/.github/workflows/build-tensorrt-linux.yml index 7581c38ae8..42fd32eb55 100644 --- a/.github/workflows/build-tensorrt-linux.yml +++ b/.github/workflows/build-tensorrt-linux.yml @@ -114,13 +114,13 @@ jobs: rm -rf "${RUNNER_TEMP}/*" fi echo "::endgroup::" - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 with: # Support the use case where we need to checkout someone's fork repository: ${{ inputs.test-infra-repository }} ref: ${{ inputs.test-infra-ref }} path: test-infra - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 if: ${{ env.ARCH == 'aarch64' }} with: # Support the use case where we need to checkout someone's fork @@ -212,7 +212,7 @@ jobs: # NB: Only upload to GitHub after passing smoke tests - name: Upload wheel to GitHub continue-on-error: true - uses: actions/upload-artifact@v3 + uses: actions/upload-artifact@v4 with: name: ${{ env.UPLOAD_ARTIFACT_NAME }} path: ${{ inputs.repository }}/dist diff --git a/.github/workflows/build-tensorrt-windows.yml b/.github/workflows/build-tensorrt-windows.yml index 4b86910768..67639a3f02 100644 --- a/.github/workflows/build-tensorrt-windows.yml +++ b/.github/workflows/build-tensorrt-windows.yml @@ -100,7 +100,7 @@ jobs: # to have a conversation timeout-minutes: 120 steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 with: # Support the use case where we need to checkout someone's fork repository: ${{ inputs.test-infra-repository }} @@ -216,7 +216,7 @@ jobs: # NB: Only upload to GitHub after passing smoke tests - name: Upload wheel to GitHub continue-on-error: true - uses: actions/upload-artifact@v3 + uses: actions/upload-artifact@v4 with: name: ${{ env.UPLOAD_ARTIFACT_NAME }} path: ${{ inputs.repository }}/dist/ diff --git a/.github/workflows/build-test-linux.yml b/.github/workflows/build-test-linux.yml index b0a487bb79..5718e96642 100644 --- a/.github/workflows/build-test-linux.yml +++ b/.github/workflows/build-test-linux.yml @@ -30,10 +30,10 @@ jobs: matrix: ${{ steps.generate.outputs.matrix }} runs-on: ubuntu-latest steps: - - uses: actions/setup-python@v4 + - uses: actions/setup-python@v5 with: python-version: '3.10' - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 with: repository: pytorch/tensorrt - name: Generate release matrix @@ -136,7 +136,9 @@ jobs: export USE_HOST_DEPS=1 export CI_BUILD=1 pushd . - cd tests/py/dynamo + cd tests/py + python -m pip install -r requirements.txt + cd dynamo python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/dynamo_converters_test_results.xml -n 4 conversion/ popd @@ -165,7 +167,9 @@ jobs: export USE_HOST_DEPS=1 export CI_BUILD=1 pushd . - cd tests/py/dynamo + cd tests/py + python -m pip install -r requirements.txt + cd dynamo python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/dyn_models_export.xml --ir dynamo models/ popd @@ -194,7 +198,9 @@ jobs: export USE_HOST_DEPS=1 export CI_BUILD=1 pushd . - cd tests/py/dynamo + cd tests/py + python -m pip install -r requirements.txt + cd dynamo python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/export_serde_test_results.xml --ir dynamo models/test_export_serde.py python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/reexport_test_results.xml --ir dynamo models/test_reexport.py popd @@ -224,7 +230,9 @@ jobs: export USE_HOST_DEPS=1 export CI_BUILD=1 pushd . - cd tests/py/dynamo + cd tests/py + python -m pip install -r requirements.txt + cd dynamo python -m pytest -ra -n 10 --junitxml=${RUNNER_TEST_RESULTS_DIR}/torch_compile_be_test_results.xml backend/ python -m pytest -ra -n 4 --junitxml=${RUNNER_TEST_RESULTS_DIR}/torch_complete_be_e2e_test_results.xml --ir torch_compile models/test_models.py python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/torch_compile_dyn_models_export.xml --ir torch_compile models/test_dyn_models.py @@ -255,7 +263,9 @@ jobs: export USE_HOST_DEPS=1 export CI_BUILD=1 pushd . - cd tests/py/dynamo + cd tests/py + python -m pip install -r requirements.txt + cd dynamo python -m pytest -ra -n 4 --junitxml=${RUNNER_TEST_RESULTS_DIR}/tests_py_dynamo_core_runtime_test_results.xml --ignore runtime/test_002_cudagraphs_py.py --ignore runtime/test_002_cudagraphs_cpp.py runtime/ python -m pytest -ra -n 4 --junitxml=${RUNNER_TEST_RESULTS_DIR}/tests_py_dynamo_core_partitioning_test_results.xml partitioning/ python -m pytest -ra -n 4 --junitxml=${RUNNER_TEST_RESULTS_DIR}/tests_py_dynamo_core_lowering_test_results.xml lowering/ @@ -286,7 +296,9 @@ jobs: export USE_HOST_DEPS=1 export CI_BUILD=1 pushd . - cd tests/py/dynamo + cd tests/py + python -m pip install -r requirements.txt + cd dynamo nvidia-smi python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/tests_py_dynamo_core_runtime_cudagraphs_cpp_test_results.xml runtime/test_002_cudagraphs_cpp.py || true python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/tests_py_dynamo_core_runtime_cudagraphs_py_test_results.xml runtime/test_002_cudagraphs_py.py || true diff --git a/.github/workflows/build-test-tensorrt-linux.yml b/.github/workflows/build-test-tensorrt-linux.yml index 625ffe9a31..f25e96c5d6 100644 --- a/.github/workflows/build-test-tensorrt-linux.yml +++ b/.github/workflows/build-test-tensorrt-linux.yml @@ -2,6 +2,8 @@ name: Build and Test Torch-TensorRT on Linux with Future TensorRT Versions on: workflow_dispatch: + schedule: + - cron: '0 0 * * 0' # Runs at 00:00 UTC every Sunday (minute hour day-of-month month-of-year day-of-week) permissions: id-token: write @@ -25,16 +27,18 @@ jobs: matrix: ${{ steps.generate.outputs.matrix }} runs-on: ubuntu-latest steps: - - uses: actions/setup-python@v4 + - uses: actions/setup-python@v5 with: python-version: '3.10' - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 with: repository: pytorch/tensorrt - name: Generate tensorrt matrix id: generate run: | set -eou pipefail + python -m pip install --upgrade pip + pip install requests MATRIX_BLOB=${{ toJSON(needs.generate-matrix.outputs.matrix) }} MATRIX_BLOB="$(python3 .github/scripts/generate-tensorrt-test-matrix.py --matrix "${MATRIX_BLOB}")" echo "${MATRIX_BLOB}" @@ -128,7 +132,9 @@ jobs: export USE_HOST_DEPS=1 export CI_BUILD=1 pushd . - cd tests/py/dynamo + cd tests/py + python -m pip install -r requirements.txt + cd dynamo python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/dynamo_converters_test_results.xml -n 4 conversion/ popd @@ -157,7 +163,9 @@ jobs: export USE_HOST_DEPS=1 export CI_BUILD=1 pushd . - cd tests/py/dynamo + cd tests/py + python -m pip install -r requirements.txt + cd dynamo python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/dyn_models_export.xml --ir dynamo models/ popd @@ -186,7 +194,9 @@ jobs: export USE_HOST_DEPS=1 export CI_BUILD=1 pushd . - cd tests/py/dynamo + cd tests/py + python -m pip install -r requirements.txt + cd dynamo python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/export_serde_test_results.xml --ir dynamo models/test_export_serde.py popd @@ -215,7 +225,9 @@ jobs: export USE_HOST_DEPS=1 export CI_BUILD=1 pushd . - cd tests/py/dynamo + cd tests/py + python -m pip install -r requirements.txt + cd dynamo python -m pytest -ra -n 10 --junitxml=${RUNNER_TEST_RESULTS_DIR}/torch_compile_be_test_results.xml backend/ python -m pytest -ra -n 4 --junitxml=${RUNNER_TEST_RESULTS_DIR}/torch_complete_be_e2e_test_results.xml --ir torch_compile models/test_models.py python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/torch_compile_dyn_models_export.xml --ir torch_compile models/test_dyn_models.py @@ -246,7 +258,9 @@ jobs: export USE_HOST_DEPS=1 export CI_BUILD=1 pushd . - cd tests/py/dynamo + cd tests/py + python -m pip install -r requirements.txt + cd dynamo python -m pytest -ra -n 4 --junitxml=${RUNNER_TEST_RESULTS_DIR}/tests_py_dynamo_core_runtime_test_results.xml --ignore runtime/test_002_cudagraphs_py.py --ignore runtime/test_002_cudagraphs_cpp.py runtime/ python -m pytest -ra -n 4 --junitxml=${RUNNER_TEST_RESULTS_DIR}/tests_py_dynamo_core_partitioning_test_results.xml partitioning/ python -m pytest -ra -n 4 --junitxml=${RUNNER_TEST_RESULTS_DIR}/tests_py_dynamo_core_lowering_test_results.xml lowering/ @@ -277,7 +291,9 @@ jobs: export USE_HOST_DEPS=1 export CI_BUILD=1 pushd . - cd tests/py/dynamo + cd tests/py + python -m pip install -r requirements.txt + cd dynamo nvidia-smi python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/tests_py_dynamo_core_runtime_cudagraphs_cpp_test_results.xml runtime/test_002_cudagraphs_cpp.py || true python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/tests_py_dynamo_core_runtime_cudagraphs_py_test_results.xml runtime/test_002_cudagraphs_py.py || true diff --git a/.github/workflows/build-test-tensorrt-windows.yml b/.github/workflows/build-test-tensorrt-windows.yml index fe812e1b9d..6df471184c 100644 --- a/.github/workflows/build-test-tensorrt-windows.yml +++ b/.github/workflows/build-test-tensorrt-windows.yml @@ -2,6 +2,8 @@ name: Build and Test Torch-TensorRT on Windows with Future TensorRT Versions on: workflow_dispatch: + schedule: + - cron: '0 0 * * 0' # Runs at 00:00 UTC every Sunday (minute hour day-of-month month-of-year day-of-week) permissions: id-token: write @@ -25,16 +27,18 @@ jobs: matrix: ${{ steps.generate.outputs.matrix }} runs-on: ubuntu-latest steps: - - uses: actions/setup-python@v4 + - uses: actions/setup-python@v5 with: python-version: '3.10' - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 with: repository: pytorch/tensorrt - name: Generate tensorrt matrix id: generate run: | set -eou pipefail + python -m pip install --upgrade pip + pip install requests MATRIX_BLOB=${{ toJSON(needs.generate-matrix.outputs.matrix) }} MATRIX_BLOB="$(python3 .github/scripts/generate-tensorrt-test-matrix.py --matrix "${MATRIX_BLOB}")" echo "${MATRIX_BLOB}" @@ -131,7 +135,9 @@ jobs: export USE_HOST_DEPS=1 export CI_BUILD=1 pushd . - cd tests/py/dynamo + cd tests/py + python -m pip install -r requirements.txt + cd dynamo python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/dynamo_converters_test_results.xml -n 4 conversion/ popd @@ -157,7 +163,9 @@ jobs: export USE_HOST_DEPS=1 export CI_BUILD=1 pushd . - cd tests/py/dynamo + cd tests/py + python -m pip install -r requirements.txt + cd dynamo python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/dyn_models_export.xml --ir dynamo models/ popd @@ -183,7 +191,9 @@ jobs: export USE_HOST_DEPS=1 export CI_BUILD=1 pushd . - cd tests/py/dynamo + cd tests/py + python -m pip install -r requirements.txt + cd dynamo python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/export_serde_test_results.xml --ir dynamo models/test_export_serde.py popd @@ -209,7 +219,9 @@ jobs: export USE_HOST_DEPS=1 export CI_BUILD=1 pushd . - cd tests/py/dynamo + cd tests/py + python -m pip install -r requirements.txt + cd dynamo python -m pytest -ra -n 10 --junitxml=${RUNNER_TEST_RESULTS_DIR}/torch_compile_be_test_results.xml backend/ python -m pytest -ra -n 4 --junitxml=${RUNNER_TEST_RESULTS_DIR}/torch_complete_be_e2e_test_results.xml --ir torch_compile models/test_models.py python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/torch_compile_dyn_models_export.xml --ir torch_compile models/test_dyn_models.py @@ -237,7 +249,9 @@ jobs: export USE_HOST_DEPS=1 export CI_BUILD=1 pushd . - cd tests/py/dynamo + cd tests/py + python -m pip install -r requirements.txt + cd dynamo python -m pytest -ra -n 4 --junitxml=${RUNNER_TEST_RESULTS_DIR}/tests_py_dynamo_core_runtime_test_results.xml --ignore runtime/test_002_cudagraphs_py.py --ignore runtime/test_002_cudagraphs_cpp.py runtime/ python -m pytest -ra -n 4 --junitxml=${RUNNER_TEST_RESULTS_DIR}/tests_py_dynamo_core_partitioning_test_results.xml partitioning/ python -m pytest -ra -n 4 --junitxml=${RUNNER_TEST_RESULTS_DIR}/tests_py_dynamo_core_lowering_test_results.xml lowering/ @@ -265,7 +279,9 @@ jobs: export USE_HOST_DEPS=1 export CI_BUILD=1 pushd . - cd tests/py/dynamo + cd tests/py + python -m pip install -r requirements.txt + cd dynamo python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/tests_py_dynamo_core_runtime_cudagraphs_cpp_test_results.xml runtime/test_002_cudagraphs_cpp.py python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/tests_py_dynamo_core_runtime_cudagraphs_py_test_results.xml runtime/test_002_cudagraphs_py.py popd diff --git a/.github/workflows/build-test-windows.yml b/.github/workflows/build-test-windows.yml index c227d14a0f..2ee31b4b74 100644 --- a/.github/workflows/build-test-windows.yml +++ b/.github/workflows/build-test-windows.yml @@ -118,7 +118,9 @@ jobs: export USE_HOST_DEPS=1 export CI_BUILD=1 pushd . - cd tests/py/dynamo + cd tests/py + python -m pip install -r requirements.txt + cd dynamo python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/dynamo_converters_test_results.xml -n 4 conversion/ popd @@ -144,7 +146,9 @@ jobs: export USE_HOST_DEPS=1 export CI_BUILD=1 pushd . - cd tests/py/dynamo + cd tests/py + python -m pip install -r requirements.txt + cd dynamo python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/dyn_models_export.xml --ir dynamo models/ popd @@ -170,7 +174,9 @@ jobs: export USE_HOST_DEPS=1 export CI_BUILD=1 pushd . - cd tests/py/dynamo + cd tests/py + python -m pip install -r requirements.txt + cd dynamo python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/export_serde_test_results.xml --ir dynamo models/test_export_serde.py python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/reexport_test_results.xml --ir dynamo models/test_reexport.py popd @@ -197,7 +203,9 @@ jobs: export USE_HOST_DEPS=1 export CI_BUILD=1 pushd . - cd tests/py/dynamo + cd tests/py + python -m pip install -r requirements.txt + cd dynamo python -m pytest -ra -n 10 --junitxml=${RUNNER_TEST_RESULTS_DIR}/torch_compile_be_test_results.xml backend/ python -m pytest -ra -n 4 --junitxml=${RUNNER_TEST_RESULTS_DIR}/torch_complete_be_e2e_test_results.xml --ir torch_compile models/test_models.py python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/torch_compile_dyn_models_export.xml --ir torch_compile models/test_dyn_models.py @@ -225,7 +233,9 @@ jobs: export USE_HOST_DEPS=1 export CI_BUILD=1 pushd . - cd tests/py/dynamo + cd tests/py + python -m pip install -r requirements.txt + cd dynamo python -m pytest -ra -n 4 --junitxml=${RUNNER_TEST_RESULTS_DIR}/tests_py_dynamo_core_runtime_test_results.xml --ignore runtime/test_002_cudagraphs_py.py --ignore runtime/test_002_cudagraphs_cpp.py runtime/ python -m pytest -ra -n 4 --junitxml=${RUNNER_TEST_RESULTS_DIR}/tests_py_dynamo_core_partitioning_test_results.xml partitioning/ python -m pytest -ra -n 4 --junitxml=${RUNNER_TEST_RESULTS_DIR}/tests_py_dynamo_core_lowering_test_results.xml lowering/ @@ -253,7 +263,9 @@ jobs: export USE_HOST_DEPS=1 export CI_BUILD=1 pushd . - cd tests/py/dynamo + cd tests/py + python -m pip install -r requirements.txt + cd dynamo python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/tests_py_dynamo_core_runtime_cudagraphs_cpp_test_results.xml runtime/test_002_cudagraphs_cpp.py python -m pytest -ra --junitxml=${RUNNER_TEST_RESULTS_DIR}/tests_py_dynamo_core_runtime_cudagraphs_py_test_results.xml runtime/test_002_cudagraphs_py.py popd diff --git a/.github/workflows/docker_builder.yml b/.github/workflows/docker_builder.yml index a978d82b6a..b991acf933 100644 --- a/.github/workflows/docker_builder.yml +++ b/.github/workflows/docker_builder.yml @@ -30,7 +30,7 @@ jobs: steps: - name: Checkout repository - uses: actions/checkout@v3 + uses: actions/checkout@v4 - name: Fix Slashes Repo Name id: fix_slashes diff --git a/.github/workflows/linter.yml b/.github/workflows/linter.yml index 6428bef8c8..464dfbdf34 100644 --- a/.github/workflows/linter.yml +++ b/.github/workflows/linter.yml @@ -26,7 +26,7 @@ jobs: with: ref: ${{ github.event.pull_request.head.sha }} - name: Set up Python 3.9 - uses: actions/setup-python@v3 + uses: actions/setup-python@v5 with: python-version: '3.9' - name: Setup env @@ -66,7 +66,7 @@ jobs: with: ref: ${{ github.event.pull_request.head.sha }} - name: Set up Python 3.9 - uses: actions/setup-python@v3 + uses: actions/setup-python@v5 with: python-version: '3.9' - name: Setup env diff --git a/.github/workflows/linux-test.yml b/.github/workflows/linux-test.yml index 6ddc601f2c..5d47e51d04 100644 --- a/.github/workflows/linux-test.yml +++ b/.github/workflows/linux-test.yml @@ -85,7 +85,7 @@ jobs: rm -rfv "${GITHUB_WORKSPACE}" mkdir -p "${GITHUB_WORKSPACE}" echo "::endgroup::" - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 with: # Support the use case where we need to checkout someone's fork repository: ${{ inputs.test-infra-repository }} @@ -114,13 +114,13 @@ jobs: script: ${{ inputs.pre-script }} - name: Download artifacts if: ${{ matrix.tensorrt == '' }} - uses: actions/download-artifact@v3 + uses: actions/download-artifact@v4 with: name: ${{ env.ARTIFACT_NAME }} path: /opt/torch-tensorrt-builds/ - name: Download artifacts if: ${{ matrix.tensorrt != '' }} - uses: actions/download-artifact@v3 + uses: actions/download-artifact@v4 with: name: ${{ env.DOWNLOAD_ARTIFACT_NAME }} path: /opt/torch-tensorrt-builds/ @@ -155,7 +155,7 @@ jobs: path: ${{ env.RUNNER_TEST_RESULTS_DIR }}/**/*.xml summary: true display-options: fEs - fail-on-empty: True + fail-on-empty: false - name: Prepare artifacts for upload working-directory: ${{ inputs.repository }} @@ -184,7 +184,7 @@ jobs: echo "upload-docs=${upload_docs}" >> "${GITHUB_OUTPUT}" - name: Upload artifacts to GitHub (if any) - uses: actions/upload-artifact@v3 + uses: actions/upload-artifact@v4 if: ${{ inputs.upload-artifact != '' }} with: name: ${{ inputs.upload-artifact }} diff --git a/.github/workflows/nightlies.yml b/.github/workflows/nightlies.yml index a0692cdafe..aac1c58f7f 100644 --- a/.github/workflows/nightlies.yml +++ b/.github/workflows/nightlies.yml @@ -11,7 +11,7 @@ jobs: environment: trigger-nightly timeout-minutes: 120 steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 with: ref: main token: ${{ secrets.GH_PYTORCHBOT_TOKEN }} diff --git a/.github/workflows/release-linux.yml b/.github/workflows/release-linux.yml index ca13b37443..25454d27e2 100644 --- a/.github/workflows/release-linux.yml +++ b/.github/workflows/release-linux.yml @@ -31,10 +31,10 @@ jobs: matrix: ${{ steps.generate.outputs.matrix }} runs-on: ubuntu-latest steps: - - uses: actions/setup-python@v4 + - uses: actions/setup-python@v5 with: python-version: '3.10' - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 with: repository: pytorch/tensorrt - name: Generate release matrix @@ -81,10 +81,10 @@ jobs: matrix: ${{ steps.generate.outputs.matrix }} runs-on: ubuntu-latest steps: - - uses: actions/setup-python@v4 + - uses: actions/setup-python@v5 with: python-version: '3.10' - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 with: repository: pytorch/tensorrt - name: Generate release matrix diff --git a/.github/workflows/release-wheel-linux.yml b/.github/workflows/release-wheel-linux.yml index 6ddd9e0306..54732378eb 100644 --- a/.github/workflows/release-wheel-linux.yml +++ b/.github/workflows/release-wheel-linux.yml @@ -114,13 +114,13 @@ jobs: rm -rf "${RUNNER_TEMP}/*" fi echo "::endgroup::" - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 with: # Support the use case where we need to checkout someone's fork repository: ${{ inputs.test-infra-repository }} ref: ${{ inputs.test-infra-ref }} path: test-infra - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 if: ${{ env.ARCH == 'aarch64' }} with: # Support the use case where we need to checkout someone's fork @@ -236,21 +236,21 @@ jobs: - name: Upload wheel to GitHub if: ${{ inputs.cxx11-tarball-release != 'true' }} continue-on-error: true - uses: actions/upload-artifact@v3 + uses: actions/upload-artifact@v4 with: name: ${{ env.ARTIFACT_NAME }} path: ${{ inputs.repository }}/release/wheel/ - name: Upload pre-cxx11 tarball to GitHub if: ${{ inputs.cxx11-tarball-release != 'true' && env.PYTHON_VERSION == '3.10' }} continue-on-error: true - uses: actions/upload-artifact@v3 + uses: actions/upload-artifact@v4 with: name: pre-cxx11-tarball-${{ env.PYTHON_VERSION }}-${{ env.CU_VERSION }} path: ${{ inputs.repository }}/release/tarball/ - name: Upload cxx11 tarball to GitHub if: ${{ inputs.cxx11-tarball-release == 'true' }} continue-on-error: true - uses: actions/upload-artifact@v3 + uses: actions/upload-artifact@v4 with: name: cxx11-tarball-${{ env.PYTHON_VERSION }}-${{ env.CU_VERSION }} path: ${{ inputs.repository }}/release/tarball/ diff --git a/.github/workflows/release-wheel-windows.yml b/.github/workflows/release-wheel-windows.yml index 6a6c993502..2ea88bce9e 100644 --- a/.github/workflows/release-wheel-windows.yml +++ b/.github/workflows/release-wheel-windows.yml @@ -90,7 +90,7 @@ jobs: # to have a conversation timeout-minutes: 120 steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 with: # Support the use case where we need to checkout someone's fork repository: ${{ inputs.test-infra-repository }} @@ -199,7 +199,7 @@ jobs: # NB: Only upload to GitHub after passing smoke tests - name: Upload wheel to GitHub continue-on-error: true - uses: actions/upload-artifact@v3 + uses: actions/upload-artifact@v4 with: name: ${{ env.ARTIFACT_NAME }} path: ${{ inputs.repository }}/dist/ diff --git a/.github/workflows/release-windows.yml b/.github/workflows/release-windows.yml index 271547cec3..c6518e1e2a 100644 --- a/.github/workflows/release-windows.yml +++ b/.github/workflows/release-windows.yml @@ -31,10 +31,10 @@ jobs: matrix: ${{ steps.generate.outputs.matrix }} runs-on: ubuntu-latest steps: - - uses: actions/setup-python@v4 + - uses: actions/setup-python@v5 with: python-version: '3.10' - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 with: repository: pytorch/tensorrt - name: Generate release matrix diff --git a/.github/workflows/windows-test.yml b/.github/workflows/windows-test.yml index 13feedfa8c..a8b27c0aa9 100644 --- a/.github/workflows/windows-test.yml +++ b/.github/workflows/windows-test.yml @@ -70,7 +70,7 @@ jobs: mkdir -p "${GITHUB_WORKSPACE}" echo "::endgroup::" - name: Checkout repository (${{ inputs.test-infra-repository }}@${{ inputs.test-infra-ref }}) - uses: actions/checkout@v3 + uses: actions/checkout@v4 with: # Support the use case where we need to checkout someone's fork repository: ${{ inputs.test-infra-repository }} @@ -105,13 +105,13 @@ jobs: is_windows: 'enabled' - name: Download artifacts if: ${{ matrix.tensorrt == '' }} - uses: actions/download-artifact@v3 + uses: actions/download-artifact@v4 with: name: ${{ env.ARTIFACT_NAME }} path: ${{ runner.temp }}/artifacts/ - name: Download artifacts if: ${{ matrix.tensorrt != '' }} - uses: actions/download-artifact@v3 + uses: actions/download-artifact@v4 with: name: ${{ env.DOWNLOAD_ARTIFACT_NAME }} path: ${{ runner.temp }}/artifacts/ diff --git a/README.md b/README.md index 5ee8dd4df7..7312ecbd5a 100644 --- a/README.md +++ b/README.md @@ -7,7 +7,7 @@ Torch-TensorRT [![Documentation](https://img.shields.io/badge/docs-master-brightgreen)](https://nvidia.github.io/Torch-TensorRT/) [![pytorch](https://img.shields.io/badge/PyTorch-2.4-green)](https://www.python.org/downloads/release/python-31013/) [![cuda](https://img.shields.io/badge/CUDA-12.4-green)](https://developer.nvidia.com/cuda-downloads) -[![trt](https://img.shields.io/badge/TensorRT-10.3.0-green)](https://github.com/nvidia/tensorrt-llm) +[![trt](https://img.shields.io/badge/TensorRT-10.7.0-green)](https://github.com/nvidia/tensorrt-llm) [![license](https://img.shields.io/badge/license-BSD--3--Clause-blue)](./LICENSE) [![linux_tests](https://github.com/pytorch/TensorRT/actions/workflows/build-test-linux.yml/badge.svg)](https://github.com/pytorch/TensorRT/actions/workflows/build-test-linux.yml) [![windows_tests](https://github.com/pytorch/TensorRT/actions/workflows/build-test-windows.yml/badge.svg)](https://github.com/pytorch/TensorRT/actions/workflows/build-test-windows.yml) @@ -119,7 +119,7 @@ These are the following dependencies used to verify the testcases. Torch-TensorR - Bazel 6.3.2 - Libtorch 2.5.0.dev (latest nightly) (built with CUDA 12.4) - CUDA 12.4 -- TensorRT 10.6.0.26 +- TensorRT 10.7.0.23 ## Deprecation Policy diff --git a/dev_dep_versions.yml b/dev_dep_versions.yml index 527b83936e..b921d558fd 100644 --- a/dev_dep_versions.yml +++ b/dev_dep_versions.yml @@ -1,2 +1,2 @@ -__cuda_version__: "12.4" -__tensorrt_version__: ">=10.3.0,<=10.6.0" +__cuda_version__: "12.6" +__tensorrt_version__: "10.7.0" diff --git a/docs/_cpp_api/classtorch__tensorrt_1_1DataType.html b/docs/_cpp_api/classtorch__tensorrt_1_1DataType.html index 63ac537bc6..d2783e8427 100644 --- a/docs/_cpp_api/classtorch__tensorrt_1_1DataType.html +++ b/docs/_cpp_api/classtorch__tensorrt_1_1DataType.html @@ -10,7 +10,7 @@ - Class DataType — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Class DataType — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

Find resources and get questions answered

- - Contributor Awards - 2023 + + Contributor Awards - 2024

Award winners announced at this year's PyTorch Conference

@@ -157,6 +157,9 @@ ExecuTorch

End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

+ + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

Find events, webinars, and podcasts

+ + Newsletter +

Stay up-to-date with the latest updates

+
@@ -219,7 +226,18 @@ Governing Board -

+
+ + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
- v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
@@ -900,7 +918,7 @@

Resources

Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -916,6 +934,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -954,6 +975,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -967,6 +991,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/classtorch__tensorrt_1_1Device_1_1DeviceType.html b/docs/_cpp_api/classtorch__tensorrt_1_1Device_1_1DeviceType.html index 4042b0c058..e764faeec8 100644 --- a/docs/_cpp_api/classtorch__tensorrt_1_1Device_1_1DeviceType.html +++ b/docs/_cpp_api/classtorch__tensorrt_1_1Device_1_1DeviceType.html @@ -10,7 +10,7 @@ - Class Device::DeviceType — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Class Device::DeviceType — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -837,7 +855,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -853,6 +871,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -891,6 +912,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -904,6 +928,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/classtorch__tensorrt_1_1TensorFormat.html b/docs/_cpp_api/classtorch__tensorrt_1_1TensorFormat.html index 418eced762..4264c28a70 100644 --- a/docs/_cpp_api/classtorch__tensorrt_1_1TensorFormat.html +++ b/docs/_cpp_api/classtorch__tensorrt_1_1TensorFormat.html @@ -10,7 +10,7 @@ - Class TensorFormat — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Class TensorFormat — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -869,7 +887,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -885,6 +903,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -923,6 +944,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -936,6 +960,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8CacheCalibrator.html b/docs/_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8CacheCalibrator.html index d5cd609dcf..2e9ecec213 100644 --- a/docs/_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8CacheCalibrator.html +++ b/docs/_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8CacheCalibrator.html @@ -10,7 +10,7 @@ - Template Class Int8CacheCalibrator — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Template Class Int8CacheCalibrator — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -833,7 +851,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -849,6 +867,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -887,6 +908,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -900,6 +924,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8Calibrator.html b/docs/_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8Calibrator.html index c3512f24d2..a6b970cdaa 100644 --- a/docs/_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8Calibrator.html +++ b/docs/_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8Calibrator.html @@ -10,7 +10,7 @@ - Template Class Int8Calibrator — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Template Class Int8Calibrator — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -843,7 +861,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -859,6 +877,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -897,6 +918,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -910,6 +934,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/define_macros_8h_1a18d295a837ac71add5578860b55e5502.html b/docs/_cpp_api/define_macros_8h_1a18d295a837ac71add5578860b55e5502.html index 929768c994..4fb5188ef4 100644 --- a/docs/_cpp_api/define_macros_8h_1a18d295a837ac71add5578860b55e5502.html +++ b/docs/_cpp_api/define_macros_8h_1a18d295a837ac71add5578860b55e5502.html @@ -10,7 +10,7 @@ - Define STR — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Define STR — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -728,7 +746,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -744,6 +762,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -782,6 +803,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -795,6 +819,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/define_macros_8h_1a282fd3c0b1c3a215148ae372070e1268.html b/docs/_cpp_api/define_macros_8h_1a282fd3c0b1c3a215148ae372070e1268.html index 7ee7b748f0..8885064aa5 100644 --- a/docs/_cpp_api/define_macros_8h_1a282fd3c0b1c3a215148ae372070e1268.html +++ b/docs/_cpp_api/define_macros_8h_1a282fd3c0b1c3a215148ae372070e1268.html @@ -10,7 +10,7 @@ - Define TORCH_TENSORRT_PATCH_VERSION — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Define TORCH_TENSORRT_PATCH_VERSION — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -728,7 +746,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -744,6 +762,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -782,6 +803,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -795,6 +819,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/define_macros_8h_1a31398a6d4d27e28817afb0f0139e909e.html b/docs/_cpp_api/define_macros_8h_1a31398a6d4d27e28817afb0f0139e909e.html index bb2af69d65..5f20176e3c 100644 --- a/docs/_cpp_api/define_macros_8h_1a31398a6d4d27e28817afb0f0139e909e.html +++ b/docs/_cpp_api/define_macros_8h_1a31398a6d4d27e28817afb0f0139e909e.html @@ -10,7 +10,7 @@ - Define TORCH_TENSORRT_MAJOR_VERSION — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Define TORCH_TENSORRT_MAJOR_VERSION — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -728,7 +746,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -744,6 +762,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -782,6 +803,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -795,6 +819,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/define_macros_8h_1a35703561b26b1a9d2738ad7d58b27827.html b/docs/_cpp_api/define_macros_8h_1a35703561b26b1a9d2738ad7d58b27827.html index 6c5c061348..a0d6c91fdc 100644 --- a/docs/_cpp_api/define_macros_8h_1a35703561b26b1a9d2738ad7d58b27827.html +++ b/docs/_cpp_api/define_macros_8h_1a35703561b26b1a9d2738ad7d58b27827.html @@ -10,7 +10,7 @@ - Define TORCH_TENSORRT_MINOR_VERSION — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Define TORCH_TENSORRT_MINOR_VERSION — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -728,7 +746,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -744,6 +762,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -782,6 +803,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -795,6 +819,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/define_macros_8h_1abd1465eb38256d3f22cc1426b23d516b.html b/docs/_cpp_api/define_macros_8h_1abd1465eb38256d3f22cc1426b23d516b.html index 7b281338cf..fee7603d19 100644 --- a/docs/_cpp_api/define_macros_8h_1abd1465eb38256d3f22cc1426b23d516b.html +++ b/docs/_cpp_api/define_macros_8h_1abd1465eb38256d3f22cc1426b23d516b.html @@ -10,7 +10,7 @@ - Define TORCHTRT_API — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Define TORCHTRT_API — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -728,7 +746,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -744,6 +762,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -782,6 +803,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -795,6 +819,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/define_macros_8h_1abe87b341f562fd1cf40b7672e4d759da.html b/docs/_cpp_api/define_macros_8h_1abe87b341f562fd1cf40b7672e4d759da.html index c477b2204b..3cdf08867f 100644 --- a/docs/_cpp_api/define_macros_8h_1abe87b341f562fd1cf40b7672e4d759da.html +++ b/docs/_cpp_api/define_macros_8h_1abe87b341f562fd1cf40b7672e4d759da.html @@ -10,7 +10,7 @@ - Define XSTR — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Define XSTR — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -728,7 +746,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -744,6 +762,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -782,6 +803,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -795,6 +819,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/define_macros_8h_1ad19939408f7be171a74a89928b36eb59.html b/docs/_cpp_api/define_macros_8h_1ad19939408f7be171a74a89928b36eb59.html index 537252bd8a..1ac93f8ace 100644 --- a/docs/_cpp_api/define_macros_8h_1ad19939408f7be171a74a89928b36eb59.html +++ b/docs/_cpp_api/define_macros_8h_1ad19939408f7be171a74a89928b36eb59.html @@ -10,7 +10,7 @@ - Define TORCHTRT_HIDDEN — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Define TORCHTRT_HIDDEN — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -728,7 +746,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -744,6 +762,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -782,6 +803,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -795,6 +819,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/define_macros_8h_1adad592a7b1b7eed529cdf6acd584c883.html b/docs/_cpp_api/define_macros_8h_1adad592a7b1b7eed529cdf6acd584c883.html index 32c4e99feb..d8f9e7fdc1 100644 --- a/docs/_cpp_api/define_macros_8h_1adad592a7b1b7eed529cdf6acd584c883.html +++ b/docs/_cpp_api/define_macros_8h_1adad592a7b1b7eed529cdf6acd584c883.html @@ -10,7 +10,7 @@ - Define TORCH_TENSORRT_VERSION — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Define TORCH_TENSORRT_VERSION — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -728,7 +746,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -744,6 +762,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -782,6 +803,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -795,6 +819,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/dir_cpp.html b/docs/_cpp_api/dir_cpp.html index fdacba9096..8a6f1ce2a5 100644 --- a/docs/_cpp_api/dir_cpp.html +++ b/docs/_cpp_api/dir_cpp.html @@ -10,7 +10,7 @@ - Directory cpp — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Directory cpp — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -711,7 +729,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -727,6 +745,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -765,6 +786,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -778,6 +802,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/dir_cpp_include.html b/docs/_cpp_api/dir_cpp_include.html index a504688572..7a9a0f6728 100644 --- a/docs/_cpp_api/dir_cpp_include.html +++ b/docs/_cpp_api/dir_cpp_include.html @@ -10,7 +10,7 @@ - Directory include — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Directory include — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -712,7 +730,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -728,6 +746,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -766,6 +787,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -779,6 +803,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/dir_cpp_include_torch_tensorrt.html b/docs/_cpp_api/dir_cpp_include_torch_tensorrt.html index 34423871f1..89a9e4dec7 100644 --- a/docs/_cpp_api/dir_cpp_include_torch_tensorrt.html +++ b/docs/_cpp_api/dir_cpp_include_torch_tensorrt.html @@ -10,7 +10,7 @@ - Directory torch_tensorrt — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Directory torch_tensorrt — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -715,7 +733,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -731,6 +749,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -769,6 +790,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -782,6 +806,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/enum_logging_8h_1a130f65408ad8cbaee060f05e8db69558.html b/docs/_cpp_api/enum_logging_8h_1a130f65408ad8cbaee060f05e8db69558.html index 07ab3e7528..c942e9d244 100644 --- a/docs/_cpp_api/enum_logging_8h_1a130f65408ad8cbaee060f05e8db69558.html +++ b/docs/_cpp_api/enum_logging_8h_1a130f65408ad8cbaee060f05e8db69558.html @@ -10,7 +10,7 @@ - Enum Level — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Enum Level — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -766,7 +784,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -782,6 +800,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -820,6 +841,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -833,6 +857,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/enum_torch__tensorrt_8h_1a3fbe5d72e4fc624dbd038853079620eb.html b/docs/_cpp_api/enum_torch__tensorrt_8h_1a3fbe5d72e4fc624dbd038853079620eb.html index c823fdb038..5f06cf3f69 100644 --- a/docs/_cpp_api/enum_torch__tensorrt_8h_1a3fbe5d72e4fc624dbd038853079620eb.html +++ b/docs/_cpp_api/enum_torch__tensorrt_8h_1a3fbe5d72e4fc624dbd038853079620eb.html @@ -10,7 +10,7 @@ - Enum EngineCapability — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Enum EngineCapability — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -745,7 +763,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -761,6 +779,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -799,6 +820,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -812,6 +836,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/file_cpp_include_torch_tensorrt_logging.h.html b/docs/_cpp_api/file_cpp_include_torch_tensorrt_logging.h.html index 4bf4ac1c63..abeb2bd0cc 100644 --- a/docs/_cpp_api/file_cpp_include_torch_tensorrt_logging.h.html +++ b/docs/_cpp_api/file_cpp_include_torch_tensorrt_logging.h.html @@ -10,7 +10,7 @@ - File logging.h — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + File logging.h — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -767,7 +785,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -783,6 +801,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -821,6 +842,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -834,6 +858,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/file_cpp_include_torch_tensorrt_macros.h.html b/docs/_cpp_api/file_cpp_include_torch_tensorrt_macros.h.html index c9317e1afd..1e16767be1 100644 --- a/docs/_cpp_api/file_cpp_include_torch_tensorrt_macros.h.html +++ b/docs/_cpp_api/file_cpp_include_torch_tensorrt_macros.h.html @@ -10,7 +10,7 @@ - File macros.h — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + File macros.h — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -753,7 +771,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -769,6 +787,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -807,6 +828,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -820,6 +844,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/file_cpp_include_torch_tensorrt_ptq.h.html b/docs/_cpp_api/file_cpp_include_torch_tensorrt_ptq.h.html index cac3e2a6c5..840f0ccdb0 100644 --- a/docs/_cpp_api/file_cpp_include_torch_tensorrt_ptq.h.html +++ b/docs/_cpp_api/file_cpp_include_torch_tensorrt_ptq.h.html @@ -10,7 +10,7 @@ - File ptq.h — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + File ptq.h — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -764,7 +782,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -780,6 +798,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -818,6 +839,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -831,6 +855,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/file_cpp_include_torch_tensorrt_torch_tensorrt.h.html b/docs/_cpp_api/file_cpp_include_torch_tensorrt_torch_tensorrt.h.html index 08ec345998..186edb9900 100644 --- a/docs/_cpp_api/file_cpp_include_torch_tensorrt_torch_tensorrt.h.html +++ b/docs/_cpp_api/file_cpp_include_torch_tensorrt_torch_tensorrt.h.html @@ -10,7 +10,7 @@ - File torch_tensorrt.h — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + File torch_tensorrt.h — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -779,7 +797,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -795,6 +813,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -833,6 +854,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -846,6 +870,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/function_logging_8h_1a0593f776f469c20469e2f729fc7861a3.html b/docs/_cpp_api/function_logging_8h_1a0593f776f469c20469e2f729fc7861a3.html index 5acd8cdbba..0404241d46 100644 --- a/docs/_cpp_api/function_logging_8h_1a0593f776f469c20469e2f729fc7861a3.html +++ b/docs/_cpp_api/function_logging_8h_1a0593f776f469c20469e2f729fc7861a3.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::logging::get_logging_prefix — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Function torch_tensorrt::logging::get_logging_prefix — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -728,7 +746,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -744,6 +762,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -782,6 +803,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -795,6 +819,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/function_logging_8h_1a0c012cb374addd90eb1f42eaec570650.html b/docs/_cpp_api/function_logging_8h_1a0c012cb374addd90eb1f42eaec570650.html index ffcbb9062f..aefe9028b4 100644 --- a/docs/_cpp_api/function_logging_8h_1a0c012cb374addd90eb1f42eaec570650.html +++ b/docs/_cpp_api/function_logging_8h_1a0c012cb374addd90eb1f42eaec570650.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::logging::get_reportable_log_level — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Function torch_tensorrt::logging::get_reportable_log_level — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -734,7 +752,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -750,6 +768,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -788,6 +809,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -801,6 +825,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/function_logging_8h_1a56e110feaaba2c3fd44bd201fd21a76a.html b/docs/_cpp_api/function_logging_8h_1a56e110feaaba2c3fd44bd201fd21a76a.html index dd9187a7f6..a805709cc0 100644 --- a/docs/_cpp_api/function_logging_8h_1a56e110feaaba2c3fd44bd201fd21a76a.html +++ b/docs/_cpp_api/function_logging_8h_1a56e110feaaba2c3fd44bd201fd21a76a.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::logging::get_is_colored_output_on — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Function torch_tensorrt::logging::get_is_colored_output_on — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -734,7 +752,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -750,6 +768,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -788,6 +809,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -801,6 +825,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/function_logging_8h_1a7cb50492421ea9de4e3db895819df6f2.html b/docs/_cpp_api/function_logging_8h_1a7cb50492421ea9de4e3db895819df6f2.html index a75f081998..c873604bd2 100644 --- a/docs/_cpp_api/function_logging_8h_1a7cb50492421ea9de4e3db895819df6f2.html +++ b/docs/_cpp_api/function_logging_8h_1a7cb50492421ea9de4e3db895819df6f2.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::logging::set_reportable_log_level — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Function torch_tensorrt::logging::set_reportable_log_level — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -734,7 +752,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -750,6 +768,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -788,6 +809,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -801,6 +825,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/function_logging_8h_1ac46ac0901cb97e3ae6e93b45f24e90b8.html b/docs/_cpp_api/function_logging_8h_1ac46ac0901cb97e3ae6e93b45f24e90b8.html index 2e77fd4788..b0e349fed3 100644 --- a/docs/_cpp_api/function_logging_8h_1ac46ac0901cb97e3ae6e93b45f24e90b8.html +++ b/docs/_cpp_api/function_logging_8h_1ac46ac0901cb97e3ae6e93b45f24e90b8.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::logging::log — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Function torch_tensorrt::logging::log — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -737,7 +755,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -753,6 +771,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -791,6 +812,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -804,6 +828,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/function_logging_8h_1ad2efd47b6c3689e58ccc595680579ae5.html b/docs/_cpp_api/function_logging_8h_1ad2efd47b6c3689e58ccc595680579ae5.html index 724d6ff0df..f2ba5b909b 100644 --- a/docs/_cpp_api/function_logging_8h_1ad2efd47b6c3689e58ccc595680579ae5.html +++ b/docs/_cpp_api/function_logging_8h_1ad2efd47b6c3689e58ccc595680579ae5.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::logging::set_is_colored_output_on — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Function torch_tensorrt::logging::set_is_colored_output_on — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -734,7 +752,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -750,6 +768,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -788,6 +809,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -801,6 +825,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/function_logging_8h_1af8f3443813315af7901903d25dd495cc.html b/docs/_cpp_api/function_logging_8h_1af8f3443813315af7901903d25dd495cc.html index 76fb086654..3e5bf3b7d8 100644 --- a/docs/_cpp_api/function_logging_8h_1af8f3443813315af7901903d25dd495cc.html +++ b/docs/_cpp_api/function_logging_8h_1af8f3443813315af7901903d25dd495cc.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::logging::set_logging_prefix — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Function torch_tensorrt::logging::set_logging_prefix — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -728,7 +746,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -744,6 +762,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -782,6 +803,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -795,6 +819,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/function_ptq_8h_1a226e3c83379d1012cde8578c1c86b16c.html b/docs/_cpp_api/function_ptq_8h_1a226e3c83379d1012cde8578c1c86b16c.html index beeffe08c2..db11c7cabf 100644 --- a/docs/_cpp_api/function_ptq_8h_1a226e3c83379d1012cde8578c1c86b16c.html +++ b/docs/_cpp_api/function_ptq_8h_1a226e3c83379d1012cde8578c1c86b16c.html @@ -10,7 +10,7 @@ - Template Function torch_tensorrt::ptq::make_int8_cache_calibrator — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Template Function torch_tensorrt::ptq::make_int8_cache_calibrator — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -743,7 +761,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -759,6 +777,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -797,6 +818,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -810,6 +834,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/function_ptq_8h_1a6186e305f47c1d94b6130ef6c7f7e178.html b/docs/_cpp_api/function_ptq_8h_1a6186e305f47c1d94b6130ef6c7f7e178.html index f73022fb3a..1115ed08fb 100644 --- a/docs/_cpp_api/function_ptq_8h_1a6186e305f47c1d94b6130ef6c7f7e178.html +++ b/docs/_cpp_api/function_ptq_8h_1a6186e305f47c1d94b6130ef6c7f7e178.html @@ -10,7 +10,7 @@ - Template Function torch_tensorrt::ptq::make_int8_calibrator — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Template Function torch_tensorrt::ptq::make_int8_calibrator — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -749,7 +767,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -765,6 +783,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -803,6 +824,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -816,6 +840,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/function_torch__tensorrt_8h_1a5b405fd3bf3c8fc2e2a54cbbab979797.html b/docs/_cpp_api/function_torch__tensorrt_8h_1a5b405fd3bf3c8fc2e2a54cbbab979797.html index da325957b8..2b1ae77ee6 100644 --- a/docs/_cpp_api/function_torch__tensorrt_8h_1a5b405fd3bf3c8fc2e2a54cbbab979797.html +++ b/docs/_cpp_api/function_torch__tensorrt_8h_1a5b405fd3bf3c8fc2e2a54cbbab979797.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::torchscript::check_method_operator_support — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Function torch_tensorrt::torchscript::check_method_operator_support — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -743,7 +761,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -759,6 +777,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -797,6 +818,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -810,6 +834,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/function_torch__tensorrt_8h_1a6e19490a08fb1553c9dd347a5ae79db9.html b/docs/_cpp_api/function_torch__tensorrt_8h_1a6e19490a08fb1553c9dd347a5ae79db9.html index a6fc4c983d..33d68ca62f 100644 --- a/docs/_cpp_api/function_torch__tensorrt_8h_1a6e19490a08fb1553c9dd347a5ae79db9.html +++ b/docs/_cpp_api/function_torch__tensorrt_8h_1a6e19490a08fb1553c9dd347a5ae79db9.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::torchscript::compile — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Function torch_tensorrt::torchscript::compile — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -743,7 +761,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -759,6 +777,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -797,6 +818,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -810,6 +834,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/function_torch__tensorrt_8h_1a81f9783517335dda877d8cfcf38987c9.html b/docs/_cpp_api/function_torch__tensorrt_8h_1a81f9783517335dda877d8cfcf38987c9.html index 7a777d8533..ab8a6e7307 100644 --- a/docs/_cpp_api/function_torch__tensorrt_8h_1a81f9783517335dda877d8cfcf38987c9.html +++ b/docs/_cpp_api/function_torch__tensorrt_8h_1a81f9783517335dda877d8cfcf38987c9.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::torchscript::embed_engine_in_new_module — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Function torch_tensorrt::torchscript::embed_engine_in_new_module — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -749,7 +767,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -765,6 +783,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -803,6 +824,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -816,6 +840,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/function_torch__tensorrt_8h_1ac4ab8313ae72c2c899ea31548b528528.html b/docs/_cpp_api/function_torch__tensorrt_8h_1ac4ab8313ae72c2c899ea31548b528528.html index 2a17485c39..bb9c500a3a 100644 --- a/docs/_cpp_api/function_torch__tensorrt_8h_1ac4ab8313ae72c2c899ea31548b528528.html +++ b/docs/_cpp_api/function_torch__tensorrt_8h_1ac4ab8313ae72c2c899ea31548b528528.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::get_build_info — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Function torch_tensorrt::get_build_info — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -734,7 +752,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -750,6 +768,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -788,6 +809,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -801,6 +825,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/function_torch__tensorrt_8h_1ad1acd06eaeaffbbcf6e7ebf426891384.html b/docs/_cpp_api/function_torch__tensorrt_8h_1ad1acd06eaeaffbbcf6e7ebf426891384.html index 7dc212627a..ec8101ee49 100644 --- a/docs/_cpp_api/function_torch__tensorrt_8h_1ad1acd06eaeaffbbcf6e7ebf426891384.html +++ b/docs/_cpp_api/function_torch__tensorrt_8h_1ad1acd06eaeaffbbcf6e7ebf426891384.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::set_device — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Function torch_tensorrt::set_device — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -734,7 +752,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -750,6 +768,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -788,6 +809,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -801,6 +825,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/function_torch__tensorrt_8h_1ad6a4ee8ca6c8f6e5519eb1128ec7f4a1.html b/docs/_cpp_api/function_torch__tensorrt_8h_1ad6a4ee8ca6c8f6e5519eb1128ec7f4a1.html index a00d267252..a3b06aa82c 100644 --- a/docs/_cpp_api/function_torch__tensorrt_8h_1ad6a4ee8ca6c8f6e5519eb1128ec7f4a1.html +++ b/docs/_cpp_api/function_torch__tensorrt_8h_1ad6a4ee8ca6c8f6e5519eb1128ec7f4a1.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::dump_build_info — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Function torch_tensorrt::dump_build_info — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -729,7 +747,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -745,6 +763,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -783,6 +804,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -796,6 +820,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/function_torch__tensorrt_8h_1ae8d56472106eeef37fbe51ff7f40c9b2.html b/docs/_cpp_api/function_torch__tensorrt_8h_1ae8d56472106eeef37fbe51ff7f40c9b2.html index 4d7957bab0..7c1bcfc4be 100644 --- a/docs/_cpp_api/function_torch__tensorrt_8h_1ae8d56472106eeef37fbe51ff7f40c9b2.html +++ b/docs/_cpp_api/function_torch__tensorrt_8h_1ae8d56472106eeef37fbe51ff7f40c9b2.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::torchscript::convert_method_to_trt_engine — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Function torch_tensorrt::torchscript::convert_method_to_trt_engine — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -743,7 +761,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -759,6 +777,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -797,6 +818,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -810,6 +834,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/namespace_torch_tensorrt.html b/docs/_cpp_api/namespace_torch_tensorrt.html index 2a0ac59a33..1eafdcfd46 100644 --- a/docs/_cpp_api/namespace_torch_tensorrt.html +++ b/docs/_cpp_api/namespace_torch_tensorrt.html @@ -10,7 +10,7 @@ - Namespace torch_tensorrt — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Namespace torch_tensorrt — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -762,7 +780,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -778,6 +796,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -816,6 +837,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -829,6 +853,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/namespace_torch_tensorrt__logging.html b/docs/_cpp_api/namespace_torch_tensorrt__logging.html index 7b1c46e7b6..64d67b3b9a 100644 --- a/docs/_cpp_api/namespace_torch_tensorrt__logging.html +++ b/docs/_cpp_api/namespace_torch_tensorrt__logging.html @@ -10,7 +10,7 @@ - Namespace torch_tensorrt::logging — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Namespace torch_tensorrt::logging — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -743,7 +761,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -759,6 +777,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -797,6 +818,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -810,6 +834,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/namespace_torch_tensorrt__ptq.html b/docs/_cpp_api/namespace_torch_tensorrt__ptq.html index 5668095283..cd0057b715 100644 --- a/docs/_cpp_api/namespace_torch_tensorrt__ptq.html +++ b/docs/_cpp_api/namespace_torch_tensorrt__ptq.html @@ -10,7 +10,7 @@ - Namespace torch_tensorrt::ptq — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Namespace torch_tensorrt::ptq — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -739,7 +757,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -755,6 +773,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -793,6 +814,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -806,6 +830,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/namespace_torch_tensorrt__torchscript.html b/docs/_cpp_api/namespace_torch_tensorrt__torchscript.html index 2093b2bb27..1710c53a94 100644 --- a/docs/_cpp_api/namespace_torch_tensorrt__torchscript.html +++ b/docs/_cpp_api/namespace_torch_tensorrt__torchscript.html @@ -10,7 +10,7 @@ - Namespace torch_tensorrt::torchscript — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Namespace torch_tensorrt::torchscript — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -740,7 +758,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -756,6 +774,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -794,6 +815,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -807,6 +831,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_logging.h.html b/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_logging.h.html index f0b910b227..870fe14ecc 100644 --- a/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_logging.h.html +++ b/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_logging.h.html @@ -10,7 +10,7 @@ - Program Listing for File logging.h — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Program Listing for File logging.h — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -743,7 +761,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -759,6 +777,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -797,6 +818,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -810,6 +834,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_macros.h.html b/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_macros.h.html index e2c6e5cec9..19cb28f603 100644 --- a/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_macros.h.html +++ b/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_macros.h.html @@ -10,7 +10,7 @@ - Program Listing for File macros.h — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Program Listing for File macros.h — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -742,7 +760,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -758,6 +776,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -796,6 +817,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -809,6 +833,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_ptq.h.html b/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_ptq.h.html index f0acfcd125..fddffd47d4 100644 --- a/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_ptq.h.html +++ b/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_ptq.h.html @@ -10,7 +10,7 @@ - Program Listing for File ptq.h — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Program Listing for File ptq.h — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -877,7 +895,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -893,6 +911,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -931,6 +952,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -944,6 +968,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_torch_tensorrt.h.html b/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_torch_tensorrt.h.html index 4caec917c5..88133ae779 100644 --- a/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_torch_tensorrt.h.html +++ b/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_torch_tensorrt.h.html @@ -10,7 +10,7 @@ - Program Listing for File torch_tensorrt.h — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Program Listing for File torch_tensorrt.h — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -1046,7 +1064,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1062,6 +1080,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1100,6 +1121,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1113,6 +1137,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/structtorch__tensorrt_1_1Device.html b/docs/_cpp_api/structtorch__tensorrt_1_1Device.html index f61282d4e4..860ac6cd8c 100644 --- a/docs/_cpp_api/structtorch__tensorrt_1_1Device.html +++ b/docs/_cpp_api/structtorch__tensorrt_1_1Device.html @@ -10,7 +10,7 @@ - Struct Device — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Struct Device — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -885,7 +903,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -901,6 +919,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -939,6 +960,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -952,6 +976,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/structtorch__tensorrt_1_1GraphInputs.html b/docs/_cpp_api/structtorch__tensorrt_1_1GraphInputs.html index 0155f945c8..cc7e53406b 100644 --- a/docs/_cpp_api/structtorch__tensorrt_1_1GraphInputs.html +++ b/docs/_cpp_api/structtorch__tensorrt_1_1GraphInputs.html @@ -10,7 +10,7 @@ - Struct GraphInputs — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Struct GraphInputs — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -743,7 +761,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -759,6 +777,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -797,6 +818,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -810,6 +834,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/structtorch__tensorrt_1_1Input.html b/docs/_cpp_api/structtorch__tensorrt_1_1Input.html index 8b36991a64..396c7f7866 100644 --- a/docs/_cpp_api/structtorch__tensorrt_1_1Input.html +++ b/docs/_cpp_api/structtorch__tensorrt_1_1Input.html @@ -10,7 +10,7 @@ - Struct Input — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Struct Input — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -1064,7 +1082,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1080,6 +1098,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1118,6 +1139,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1131,6 +1155,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/structtorch__tensorrt_1_1torchscript_1_1CompileSpec.html b/docs/_cpp_api/structtorch__tensorrt_1_1torchscript_1_1CompileSpec.html index dfd5af8b75..b5078e37da 100644 --- a/docs/_cpp_api/structtorch__tensorrt_1_1torchscript_1_1CompileSpec.html +++ b/docs/_cpp_api/structtorch__tensorrt_1_1torchscript_1_1CompileSpec.html @@ -10,7 +10,7 @@ - Struct CompileSpec — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Struct CompileSpec — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -903,7 +921,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -919,6 +937,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -957,6 +978,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -970,6 +994,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/torch_tensort_cpp.html b/docs/_cpp_api/torch_tensort_cpp.html index c2facdd71e..2b8112a72e 100644 --- a/docs/_cpp_api/torch_tensort_cpp.html +++ b/docs/_cpp_api/torch_tensort_cpp.html @@ -10,7 +10,7 @@ - Torch-TensorRT C++ API — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Torch-TensorRT C++ API — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -1105,7 +1123,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1121,6 +1139,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1159,6 +1180,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1172,6 +1196,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_cpp_api/unabridged_orphan.html b/docs/_cpp_api/unabridged_orphan.html index 951c6f07e0..fd73b94aff 100644 --- a/docs/_cpp_api/unabridged_orphan.html +++ b/docs/_cpp_api/unabridged_orphan.html @@ -10,7 +10,7 @@ - Full API — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Full API — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -800,7 +818,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -816,6 +834,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -854,6 +875,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -867,6 +891,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_downloads/6a6052d9668b2cb8332d349d328e21c1/_rendered_examples_jupyter.zip b/docs/_downloads/6a6052d9668b2cb8332d349d328e21c1/_rendered_examples_jupyter.zip index 2340a51da2..4362ad5e1b 100644 Binary files a/docs/_downloads/6a6052d9668b2cb8332d349d328e21c1/_rendered_examples_jupyter.zip and b/docs/_downloads/6a6052d9668b2cb8332d349d328e21c1/_rendered_examples_jupyter.zip differ diff --git a/docs/_downloads/798cda8f83bd9f5e2cc93f329a04332c/_rendered_examples_python.zip b/docs/_downloads/798cda8f83bd9f5e2cc93f329a04332c/_rendered_examples_python.zip index 61bb575bf3..048c7ba512 100644 Binary files a/docs/_downloads/798cda8f83bd9f5e2cc93f329a04332c/_rendered_examples_python.zip and b/docs/_downloads/798cda8f83bd9f5e2cc93f329a04332c/_rendered_examples_python.zip differ diff --git a/docs/_modules/index.html b/docs/_modules/index.html index 3e79871679..c7d7948f08 100644 --- a/docs/_modules/index.html +++ b/docs/_modules/index.html @@ -9,7 +9,7 @@ - Overview: module code — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Overview: module code — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -132,8 +132,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -154,6 +154,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -201,6 +204,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -216,7 +223,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -272,7 +290,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -712,7 +730,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -728,6 +746,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -766,6 +787,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -779,6 +803,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_modules/torch_tensorrt/_Device.html b/docs/_modules/torch_tensorrt/_Device.html index ed38be75c7..06eddd6fb6 100644 --- a/docs/_modules/torch_tensorrt/_Device.html +++ b/docs/_modules/torch_tensorrt/_Device.html @@ -9,7 +9,7 @@ - torch_tensorrt._Device — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt._Device — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -132,8 +132,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -154,6 +154,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -201,6 +204,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -216,7 +223,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -272,7 +290,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -467,25 +485,25 @@

    Source code for torch_tensorrt._Device

    -from __future__ import annotations
    +from __future__ import annotations
     
    -import logging
    -import sys
    -from typing import Any, Optional, Tuple
    +import logging
    +import sys
    +from typing import Any, Optional, Tuple
     
     if sys.version_info >= (3, 11):
    -    from typing import Self
    +    from typing import Self
     else:
    -    from typing_extensions import Self
    +    from typing_extensions import Self
     
    -import torch
    -from torch_tensorrt._enums import DeviceType
    -from torch_tensorrt._features import needs_torch_tensorrt_runtime
    +import torch
    +from torch_tensorrt._enums import DeviceType
    +from torch_tensorrt._features import needs_torch_tensorrt_runtime
     
    -import tensorrt as trt
    +import tensorrt as trt
     
     
    -
    [docs]class Device(object): +
    [docs]class Device(object): """ Defines a device that can be used to specify target devices for engines @@ -505,7 +523,7 @@

    Source code for torch_tensorrt._Device

             False  #: Whether falling back to GPU if DLA cannot support an op should be allowed
         )
     
    -
    [docs] def __init__(self, *args: Any, **kwargs: Any): +
    [docs] def __init__(self, *args: Any, **kwargs: Any): """__init__ Method for torch_tensorrt.Device Device accepts one of a few construction patterns @@ -577,7 +595,7 @@

    Source code for torch_tensorrt._Device

                 if isinstance(kwargs["device_type"], trt.DeviceType):
                     self.device_type = DeviceType._from(kwargs["device_type"])
    - def __str__(self) -> str: + def __str__(self) -> str: suffix = ( ")" if self.device_type == DeviceType.GPU @@ -586,11 +604,11 @@

    Source code for torch_tensorrt._Device

             dev_str: str = f"Device(type={self.device_type}, gpu_id={self.gpu_id}{suffix}"
             return dev_str
     
    -    def __repr__(self) -> str:
    +    def __repr__(self) -> str:
             return self.__str__()
     
         @classmethod
    -    def _from(cls, d: Optional[Self | torch.device | str]) -> Device:
    +    def _from(cls, d: Optional[Self | torch.device | str]) -> Device:
             """Cast a device-type to torch_tensorrt.Device
     
             Returns the corresponding torch_tensorrt.Device
    @@ -610,16 +628,16 @@ 

    Source code for torch_tensorrt._Device

                 return cls(d)
     
         @classmethod
    -    def _from_torch_device(cls, torch_dev: torch.device) -> Device:
    +    def _from_torch_device(cls, torch_dev: torch.device) -> Device:
             return cls._from(torch_dev)
     
         @classmethod
    -    def _current_device(cls) -> Device:
    +    def _current_device(cls) -> Device:
             dev_id = torch.cuda.current_device()
             return cls(gpu_id=dev_id)
     
         @staticmethod
    -    def _parse_device_str(s: str) -> Tuple[trt.DeviceType, int]:
    +    def _parse_device_str(s: str) -> Tuple[trt.DeviceType, int]:
             s = s.lower()
             spec = s.split(":")
             if spec[0] == "gpu" or spec[0] == "cuda":
    @@ -629,7 +647,7 @@ 

    Source code for torch_tensorrt._Device

             else:
                 raise ValueError(f"Unknown device type {spec[0]}")
     
    -    def to(self, t: type) -> torch.device:
    +    def to(self, t: type) -> torch.device:
             if t == torch.device:
                 if self.gpu_id != -1:
                     return torch.device(self.gpu_id)
    @@ -639,7 +657,7 @@ 

    Source code for torch_tensorrt._Device

                 raise TypeError("Unsupported target type for device conversion")
     
         @needs_torch_tensorrt_runtime
    -    def _to_serialized_rt_device(self) -> str:
    +    def _to_serialized_rt_device(self) -> str:
             delim = torch.ops.tensorrt.SERIALIZED_RT_DEVICE_DELIM()[0]
             dev_info = torch.cuda.get_device_properties(self.gpu_id)
             rt_info = [
    @@ -879,7 +897,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -895,6 +913,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -933,6 +954,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -946,6 +970,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_modules/torch_tensorrt/_Input.html b/docs/_modules/torch_tensorrt/_Input.html index d4140ef3db..b5478c5ce6 100644 --- a/docs/_modules/torch_tensorrt/_Input.html +++ b/docs/_modules/torch_tensorrt/_Input.html @@ -9,7 +9,7 @@ - torch_tensorrt._Input — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt._Input — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -132,8 +132,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -154,6 +154,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -201,6 +204,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -216,7 +223,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -272,7 +290,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -467,16 +485,16 @@

    Source code for torch_tensorrt._Input

    -from __future__ import annotations
    +from __future__ import annotations
     
    -from enum import Enum
    -from typing import Any, Dict, List, Optional, Sequence, Tuple
    +from enum import Enum
    +from typing import Any, Dict, List, Optional, Sequence, Tuple
     
    -import torch
    -from torch_tensorrt._enums import dtype, memory_format
    +import torch
    +from torch_tensorrt._enums import dtype, memory_format
     
     
    -
    [docs]class Input(object): +
    [docs]class Input(object): """ Defines an input to a module in terms of expected shape, data type and tensor format. @@ -493,7 +511,7 @@

    Source code for torch_tensorrt._Input

             format (torch_tensorrt.TensorFormat): The expected format of the input tensor (default: torch_tensorrt.TensorFormat.NCHW)
         """
     
    -    class _ShapeMode(Enum):
    +    class _ShapeMode(Enum):
             STATIC = 0
             DYNAMIC = 1
     
    @@ -518,7 +536,7 @@ 

    Source code for torch_tensorrt._Input

         name: str = ""
         is_shape_tensor: bool = False
     
    -
    [docs] def __init__(self, *args: Any, **kwargs: Any) -> None: +
    [docs] def __init__(self, *args: Any, **kwargs: Any) -> None: """__init__ Method for torch_tensorrt.Input Input accepts one of a few construction patterns @@ -659,7 +677,7 @@

    Source code for torch_tensorrt._Input

             if "name" in kwargs:
                 self.name = kwargs["name"]
    - def __str__(self) -> str: + def __str__(self) -> str: if self.shape_mode == Input._ShapeMode.STATIC: return "Input(shape={}, dtype={}, format={}, domain=[{}, {}))".format( self.shape, @@ -686,11 +704,11 @@

    Source code for torch_tensorrt._Input

             else:
                 raise RuntimeError("Unknown input shape mode")
     
    -    def __repr__(self) -> str:
    +    def __repr__(self) -> str:
             return self.__str__()
     
         @staticmethod
    -    def equivalent_spec(a: Input, b: Input) -> bool:
    +    def equivalent_spec(a: Input, b: Input) -> bool:
             if a.shape_mode != b.shape_mode:
                 return False
     
    @@ -718,7 +736,7 @@ 

    Source code for torch_tensorrt._Input

                 return all(checks)
     
         @staticmethod
    -    def _supported_input_size_type(input_size: Any) -> bool:
    +    def _supported_input_size_type(input_size: Any) -> bool:
             if isinstance(input_size, torch.Size):
                 return True
             elif isinstance(input_size, tuple):
    @@ -729,7 +747,7 @@ 

    Source code for torch_tensorrt._Input

                 return False
     
         @staticmethod
    -    def _parse_tensor_domain(
    +    def _parse_tensor_domain(
             domain: Optional[Tuple[float, float]]
         ) -> Tuple[float, float]:
             """
    @@ -777,7 +795,7 @@ 

    Source code for torch_tensorrt._Input

             return result_domain
     
     
    [docs] @classmethod - def from_tensor( + def from_tensor( cls, t: torch.Tensor, disable_memory_format_check: bool = False ) -> "Input": """ @@ -809,7 +827,7 @@

    Source code for torch_tensorrt._Input

             return cls(shape=t.shape, dtype=t.dtype, format=frmt, torch_tensor=t)
    [docs] @classmethod - def from_tensors( + def from_tensors( cls, ts: Sequence[torch.Tensor], disable_memory_format_check: bool = False ) -> List["Input"]: """ @@ -830,7 +848,7 @@

    Source code for torch_tensorrt._Input

                 for t in ts
             ]
    -
    [docs] def example_tensor( +
    [docs] def example_tensor( self, optimization_profile_field: Optional[str] = None ) -> torch.Tensor: """ @@ -1108,7 +1126,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1124,6 +1142,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1162,6 +1183,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1175,6 +1199,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_modules/torch_tensorrt/_compile.html b/docs/_modules/torch_tensorrt/_compile.html index 045f13299d..93cf4a33ae 100644 --- a/docs/_modules/torch_tensorrt/_compile.html +++ b/docs/_modules/torch_tensorrt/_compile.html @@ -9,7 +9,7 @@ - torch_tensorrt._compile — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt._compile — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -132,8 +132,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -154,6 +154,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -201,6 +204,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -216,7 +223,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -272,7 +290,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -467,51 +485,51 @@

    Source code for torch_tensorrt._compile

    -from __future__ import annotations
    -
    -import collections.abc
    -import logging
    -import platform
    -from enum import Enum
    -from typing import Any, Callable, List, Optional, Sequence, Set
    -
    -import torch
    -import torch.fx
    -from torch_tensorrt._enums import dtype
    -from torch_tensorrt._features import ENABLED_FEATURES
    -from torch_tensorrt._Input import Input
    -from torch_tensorrt.dynamo import _defaults
    -from torch_tensorrt.dynamo.runtime._CudaGraphsTorchTensorRTModule import (
    +from __future__ import annotations
    +
    +import collections.abc
    +import logging
    +import platform
    +from enum import Enum
    +from typing import Any, Callable, List, Optional, Sequence, Set
    +
    +import torch
    +import torch.fx
    +from torch_tensorrt._enums import dtype
    +from torch_tensorrt._features import ENABLED_FEATURES
    +from torch_tensorrt._Input import Input
    +from torch_tensorrt.dynamo import _defaults
    +from torch_tensorrt.dynamo.runtime._CudaGraphsTorchTensorRTModule import (
         CudaGraphsTorchTensorRTModule,
     )
    -from torch_tensorrt.fx import InputTensorSpec
    -from torch_tensorrt.fx.lower import compile as fx_compile
    -from torch_tensorrt.fx.utils import LowerPrecision
    -from typing_extensions import TypeGuard
    +from torch_tensorrt.fx import InputTensorSpec
    +from torch_tensorrt.fx.lower import compile as fx_compile
    +from torch_tensorrt.fx.utils import LowerPrecision
    +from typing_extensions import TypeGuard
     
     if ENABLED_FEATURES.torchscript_frontend:
    -    import torch_tensorrt.ts
    -    from torch_tensorrt.ts._compiler import compile as torchscript_compile
    -    from torch_tensorrt.ts._compiler import (
    +    import torch_tensorrt.ts
    +    from torch_tensorrt.ts._compiler import compile as torchscript_compile
    +    from torch_tensorrt.ts._compiler import (
             convert_method_to_trt_engine as ts_convert_method_to_trt_engine,
         )
     
     if ENABLED_FEATURES.dynamo_frontend:
    -    from torch.export import ExportedProgram
    -    from torch_tensorrt.dynamo._compiler import compile as dynamo_compile
    -    from torch_tensorrt.dynamo._compiler import (
    +    from torch.export import ExportedProgram
    +    from torch_tensorrt.dynamo._compiler import compile as dynamo_compile
    +    from torch_tensorrt.dynamo._compiler import (
             convert_exported_program_to_serialized_trt_engine as dynamo_convert_exported_program_to_serialized_trt_engine,
         )
    -    from torch_tensorrt.dynamo._compiler import (
    +    from torch_tensorrt.dynamo._compiler import (
             cross_compile_for_windows as dynamo_cross_compile_for_windows,
         )
    -    from torch_tensorrt.dynamo._compiler import (
    +    from torch_tensorrt.dynamo._compiler import (
             load_cross_compiled_exported_program as dynamo_load_cross_compiled_exported_program,
         )
    -    from torch_tensorrt.dynamo._compiler import (
    +    from torch_tensorrt.dynamo._compiler import (
             save_cross_compiled_exported_program as dynamo_save_cross_compiled_exported_program,
         )
    -    from torch_tensorrt.dynamo._tracer import trace as dynamo_trace
    +    from torch_tensorrt.dynamo._tracer import trace as dynamo_trace
     
     logger = logging.getLogger(__name__)
     
    @@ -525,19 +543,19 @@ 

    Source code for torch_tensorrt._compile

     ]
     
     
    -def _non_fx_input_interface(
    +def _non_fx_input_interface(
         inputs: Sequence[Input | torch.Tensor | InputTensorSpec],
     ) -> TypeGuard[List[Input | torch.Tensor]]:
         return all(isinstance(i, (torch.Tensor, Input)) for i in inputs)
     
     
    -def _fx_input_interface(
    +def _fx_input_interface(
         inputs: Sequence[Input | torch.Tensor | InputTensorSpec],
     ) -> TypeGuard[List[InputTensorSpec | torch.Tensor]]:
         return all(isinstance(i, (torch.Tensor, InputTensorSpec)) for i in inputs)
     
     
    -class _IRType(Enum):
    +class _IRType(Enum):
         """Enum to determine the type of IR selected for model compilation"""
     
         ts = 0
    @@ -547,7 +565,7 @@ 

    Source code for torch_tensorrt._compile

         exported_program = 4
     
     
    -class _ModuleType(Enum):
    +class _ModuleType(Enum):
         """Enum to determine the type of model provided as input"""
     
         nn = 0
    @@ -556,7 +574,7 @@ 

    Source code for torch_tensorrt._compile

         ep = 3
     
     
    -def _parse_module_type(module: Any) -> _ModuleType:
    +def _parse_module_type(module: Any) -> _ModuleType:
         if any(
             isinstance(module, t)
             for t in [torch.jit.ScriptModule, torch.jit.ScriptFunction]
    @@ -572,7 +590,7 @@ 

    Source code for torch_tensorrt._compile

             raise RuntimeError("Module is an unknown format")
     
     
    -def _get_target_fe(module_type: _ModuleType, ir: str) -> _IRType:
    +def _get_target_fe(module_type: _ModuleType, ir: str) -> _IRType:
         module_is_tsable = any(module_type == t for t in [_ModuleType.nn, _ModuleType.ts])
         module_is_fxable = any(module_type == t for t in [_ModuleType.nn, _ModuleType.fx])
         module_is_exportable = module_type == _ModuleType.ep
    @@ -633,7 +651,7 @@ 

    Source code for torch_tensorrt._compile

                 raise ValueError("Unknown ir was requested")
     
     
    -
    [docs]def compile( +
    [docs]def compile( module: Any, ir: str = "default", inputs: Optional[Sequence[Input | torch.Tensor | InputTensorSpec]] = None, @@ -743,7 +761,7 @@

    Source code for torch_tensorrt._compile

             if kwarg_inputs is None:
                 kwarg_inputs = {}
     
    -        from torch_tensorrt.dynamo.utils import prepare_inputs
    +        from torch_tensorrt.dynamo.utils import prepare_inputs
     
             if not isinstance(arg_inputs, collections.abc.Sequence):
                 arg_inputs = [arg_inputs]  # type: ignore
    @@ -770,7 +788,7 @@ 

    Source code for torch_tensorrt._compile

             raise RuntimeError("Module is an unknown format or the ir requested is unknown")
    -def cross_compile_for_windows( +def cross_compile_for_windows( module: torch.nn.Module, file_path: str, inputs: Optional[Sequence[Input | torch.Tensor]] = None, @@ -843,7 +861,7 @@

    Source code for torch_tensorrt._compile

         if kwarg_inputs is None:
             kwarg_inputs = {}
     
    -    from torch_tensorrt.dynamo.utils import prepare_inputs
    +    from torch_tensorrt.dynamo.utils import prepare_inputs
     
         if not isinstance(arg_inputs, collections.abc.Sequence):
             arg_inputs = [arg_inputs]  # type: ignore
    @@ -869,13 +887,13 @@ 

    Source code for torch_tensorrt._compile

         logger.debug("successfully compiled and saved the module for windows")
     
     
    -def torch_compile(module: torch.nn.Module, **kwargs: Any) -> Any:
    +def torch_compile(module: torch.nn.Module, **kwargs: Any) -> Any:
         """
         Returns a boxed model which is the output of torch.compile.
         This does not compile the model to TRT. Execute this model on
         sample inputs to compile the model to TRT.
         """
    -    from torch_tensorrt.dynamo.backend import torch_tensorrt_backend
    +    from torch_tensorrt.dynamo.backend import torch_tensorrt_backend
     
         # TODO: Remove dynamic=False when SymInt Dynamic shape support is ready
         boxed_fn = torch.compile(
    @@ -885,7 +903,7 @@ 

    Source code for torch_tensorrt._compile

         return boxed_fn
     
     
    -
    [docs]def convert_method_to_trt_engine( +
    [docs]def convert_method_to_trt_engine( module: Any, method_name: str = "forward", inputs: Optional[Sequence[Input | torch.Tensor | InputTensorSpec]] = None, @@ -966,7 +984,7 @@

    Source code for torch_tensorrt._compile

             if kwarg_inputs is None:
                 kwarg_inputs = {}
     
    -        from torch_tensorrt.dynamo.utils import prepare_inputs
    +        from torch_tensorrt.dynamo.utils import prepare_inputs
     
             if not isinstance(arg_inputs, collections.abc.Sequence):
                 arg_inputs = [arg_inputs]  # type: ignore
    @@ -994,7 +1012,7 @@ 

    Source code for torch_tensorrt._compile

             raise RuntimeError("Module is an unknown format or the ir requested is unknown")
    -def load_cross_compiled_exported_program(file_path: str = "") -> Any: +def load_cross_compiled_exported_program(file_path: str = "") -> Any: """ Load an ExportedProgram file in Windows which was previously cross compiled in Linux @@ -1007,7 +1025,7 @@

    Source code for torch_tensorrt._compile

         return dynamo_load_cross_compiled_exported_program(file_path)
     
     
    -
    [docs]def load(file_path: str = "") -> Any: +
    [docs]def load(file_path: str = "") -> Any: """ Load either a Torchscript model or ExportedProgram. @@ -1044,7 +1062,7 @@

    Source code for torch_tensorrt._compile

             )
    -
    [docs]def save( +
    [docs]def save( module: Any, file_path: str = "", *, @@ -1131,7 +1149,7 @@

    Source code for torch_tensorrt._compile

                 torch.jit.save(module_ts, file_path)
             else:
                 if not retrace:
    -                from torch_tensorrt.dynamo._exporter import export
    +                from torch_tensorrt.dynamo._exporter import export
     
                     if arg_inputs is not None:
                         logger.warning(
    @@ -1378,7 +1396,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1394,6 +1412,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1432,6 +1453,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1445,6 +1469,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_modules/torch_tensorrt/_enums.html b/docs/_modules/torch_tensorrt/_enums.html index cc622fcf66..97f42a78e6 100644 --- a/docs/_modules/torch_tensorrt/_enums.html +++ b/docs/_modules/torch_tensorrt/_enums.html @@ -9,7 +9,7 @@ - torch_tensorrt._enums — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt._enums — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -132,8 +132,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -154,6 +154,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -201,6 +204,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -216,7 +223,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -272,7 +290,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -467,19 +485,19 @@

    Source code for torch_tensorrt._enums

    -from __future__ import annotations
    +from __future__ import annotations
     
    -import logging
    -from enum import Enum, auto
    -from typing import Any, Optional, Type, Union
    +import logging
    +from enum import Enum, auto
    +from typing import Any, Optional, Type, Union
     
    -import numpy as np
    -import tensorrt as trt
    -import torch
    -from torch_tensorrt._features import ENABLED_FEATURES, needs_torch_tensorrt_runtime
    +import numpy as np
    +import tensorrt as trt
    +import torch
    +from torch_tensorrt._features import ENABLED_FEATURES, needs_torch_tensorrt_runtime
     
     
    -
    [docs]class dtype(Enum): +
    [docs]class dtype(Enum): """Enum to describe data types to Torch-TensorRT, has compatibility with torch, tensorrt and numpy dtypes""" # Supported types in Torch-TensorRT @@ -575,7 +593,7 @@

    Source code for torch_tensorrt._enums

         bfloat16 = bf16
     
         @staticmethod
    -    def _is_np_obj(t: Any) -> bool:
    +    def _is_np_obj(t: Any) -> bool:
             if isinstance(t, np.dtype):
                 return True
             elif isinstance(t, type):
    @@ -584,7 +602,7 @@ 

    Source code for torch_tensorrt._enums

             return False
     
         @classmethod
    -    def _from(
    +    def _from(
             cls,
             t: Union[torch.dtype, trt.DataType, np.dtype, dtype, type],
             use_default: bool = False,
    @@ -710,7 +728,7 @@ 

    Source code for torch_tensorrt._enums

                 return t
     
             elif ENABLED_FEATURES.torchscript_frontend:
    -            from torch_tensorrt import _C
    +            from torch_tensorrt import _C
     
                 if isinstance(t, _C.dtype):
                     if t == _C.dtype.long:
    @@ -739,7 +757,7 @@ 

    Source code for torch_tensorrt._enums

             )
     
     
    [docs] @classmethod - def try_from( + def try_from( cls, t: Union[torch.dtype, trt.DataType, np.dtype, dtype], use_default: bool = False, @@ -779,7 +797,7 @@

    Source code for torch_tensorrt._enums

                 )
                 return None
    -
    [docs] def to( +
    [docs] def to( self, t: Union[Type[torch.dtype], Type[trt.DataType], Type[np.dtype], Type[dtype]], use_default: bool = False, @@ -898,7 +916,7 @@

    Source code for torch_tensorrt._enums

                 return self
     
             elif ENABLED_FEATURES.torchscript_frontend:
    -            from torch_tensorrt import _C
    +            from torch_tensorrt import _C
     
                 if t == _C.dtype:
                     if self == dtype.i64:
    @@ -926,7 +944,7 @@ 

    Source code for torch_tensorrt._enums

                 f"Provided unsupported destination type for dtype conversion {t}"
             )
    -
    [docs] def try_to( +
    [docs] def try_to( self, t: Union[Type[torch.dtype], Type[trt.DataType], Type[np.dtype], Type[dtype]], use_default: bool, @@ -965,11 +983,11 @@

    Source code for torch_tensorrt._enums

                 )
                 return None
    - def __eq__(self, other: Union[torch.dtype, trt.DataType, np.dtype, dtype]) -> bool: + def __eq__(self, other: Union[torch.dtype, trt.DataType, np.dtype, dtype]) -> bool: other_ = dtype._from(other) return bool(self.value == other_.value) - def __hash__(self) -> int: + def __hash__(self) -> int: return hash(self.value) # Putting aliases here that mess with mypy @@ -977,7 +995,7 @@

    Source code for torch_tensorrt._enums

         int = i32
    -
    [docs]class memory_format(Enum): +
    [docs]class memory_format(Enum): """""" # TensorRT supported memory layouts @@ -1109,7 +1127,7 @@

    Source code for torch_tensorrt._enums

         channels_last_3d = dhwc
     
         @classmethod
    -    def _from(
    +    def _from(
             cls, f: Union[torch.memory_format, trt.TensorFormat, memory_format]
         ) -> memory_format:
             """Create a Torch-TensorRT memory format enum from another library memory format enum.
    @@ -1185,7 +1203,7 @@ 

    Source code for torch_tensorrt._enums

                 return f
     
             elif ENABLED_FEATURES.torchscript_frontend:
    -            from torch_tensorrt import _C
    +            from torch_tensorrt import _C
     
                 if isinstance(f, _C.TensorFormat):
                     if f == _C.TensorFormat.contiguous:
    @@ -1200,7 +1218,7 @@ 

    Source code for torch_tensorrt._enums

             raise TypeError("Provided unsupported source type for memory_format conversion")
     
     
    [docs] @classmethod - def try_from( + def try_from( cls, f: Union[torch.memory_format, trt.TensorFormat, memory_format] ) -> Optional[memory_format]: """Create a Torch-TensorRT memory format enum from another library memory format enum. @@ -1233,7 +1251,7 @@

    Source code for torch_tensorrt._enums

                 )
                 return None
    -
    [docs] def to( +
    [docs] def to( self, t: Union[ Type[torch.memory_format], Type[trt.TensorFormat], Type[memory_format] @@ -1308,7 +1326,7 @@

    Source code for torch_tensorrt._enums

                 return self
     
             elif ENABLED_FEATURES.torchscript_frontend:
    -            from torch_tensorrt import _C
    +            from torch_tensorrt import _C
     
                 if t == _C.TensorFormat:
                     if self == memory_format.contiguous:
    @@ -1324,7 +1342,7 @@ 

    Source code for torch_tensorrt._enums

                 "Provided unsupported destination type for memory format conversion"
             )
    -
    [docs] def try_to( +
    [docs] def try_to( self, t: Union[ Type[torch.memory_format], Type[trt.TensorFormat], Type[memory_format] @@ -1359,17 +1377,17 @@

    Source code for torch_tensorrt._enums

                 )
                 return None
    - def __eq__( + def __eq__( self, other: Union[torch.memory_format, trt.TensorFormat, memory_format] ) -> bool: other_ = memory_format._from(other) return self.value == other_.value - def __hash__(self) -> int: + def __hash__(self) -> int: return hash(self.value)
    -
    [docs]class DeviceType(Enum): +
    [docs]class DeviceType(Enum): """Type of device TensorRT will target""" UNKNOWN = auto() @@ -1394,7 +1412,7 @@

    Source code for torch_tensorrt._enums

         """
     
         @classmethod
    -    def _from(cls, d: Union[trt.DeviceType, DeviceType]) -> DeviceType:
    +    def _from(cls, d: Union[trt.DeviceType, DeviceType]) -> DeviceType:
             """Create a Torch-TensorRT device type enum from a TensorRT device type enum.
     
             Takes a device type enum from tensorrt and create a ``torch_tensorrt.DeviceType``.
    @@ -1433,7 +1451,7 @@ 

    Source code for torch_tensorrt._enums

                 return d
     
             elif ENABLED_FEATURES.torchscript_frontend:
    -            from torch_tensorrt import _C
    +            from torch_tensorrt import _C
     
                 if isinstance(d, _C.DeviceType):
                     if d == _C.DeviceType.GPU:
    @@ -1448,7 +1466,7 @@ 

    Source code for torch_tensorrt._enums

             raise TypeError("Provided unsupported source type for DeviceType conversion")
     
     
    [docs] @classmethod - def try_from(cls, d: Union[trt.DeviceType, DeviceType]) -> Optional[DeviceType]: + def try_from(cls, d: Union[trt.DeviceType, DeviceType]) -> Optional[DeviceType]: """Create a Torch-TensorRT device type enum from a TensorRT device type enum. Takes a device type enum from tensorrt and create a ``torch_tensorrt.DeviceType``. @@ -1480,7 +1498,7 @@

    Source code for torch_tensorrt._enums

                 )
                 return None
    -
    [docs] def to( +
    [docs] def to( self, t: Union[Type[trt.DeviceType], Type[DeviceType]], use_default: bool = False, @@ -1526,7 +1544,7 @@

    Source code for torch_tensorrt._enums

                 return self
     
             elif ENABLED_FEATURES.torchscript_frontend:
    -            from torch_tensorrt import _C
    +            from torch_tensorrt import _C
     
                 if t == _C.DeviceType:
                     if self == DeviceType.GPU:
    @@ -1542,7 +1560,7 @@ 

    Source code for torch_tensorrt._enums

                 "Provided unsupported destination type for device type conversion"
             )
    -
    [docs] def try_to( +
    [docs] def try_to( self, t: Union[Type[trt.DeviceType], Type[DeviceType]], use_default: bool = False, @@ -1575,15 +1593,15 @@

    Source code for torch_tensorrt._enums

                 )
                 return None
    - def __eq__(self, other: Union[trt.DeviceType, DeviceType]) -> bool: + def __eq__(self, other: Union[trt.DeviceType, DeviceType]) -> bool: other_ = DeviceType._from(other) return bool(self.value == other_.value) - def __hash__(self) -> int: + def __hash__(self) -> int: return hash(self.value)
    -
    [docs]class EngineCapability(Enum): +
    [docs]class EngineCapability(Enum): """ EngineCapability determines the restrictions of a network during build time and what runtime it targets. """ @@ -1610,7 +1628,7 @@

    Source code for torch_tensorrt._enums

         """
     
         @classmethod
    -    def _from(
    +    def _from(
             cls, c: Union[trt.EngineCapability, EngineCapability]
         ) -> EngineCapability:
             """Create a Torch-TensorRT Engine capability enum from a TensorRT Engine capability enum.
    @@ -1651,7 +1669,7 @@ 

    Source code for torch_tensorrt._enums

                 return c
     
             elif ENABLED_FEATURES.torchscript_frontend:
    -            from torch_tensorrt import _C
    +            from torch_tensorrt import _C
     
                 if isinstance(c, _C.EngineCapability):
                     if c == _C.EngineCapability.STANDARD:
    @@ -1668,7 +1686,7 @@ 

    Source code for torch_tensorrt._enums

             )
     
     
    [docs] @classmethod - def try_from( + def try_from( c: Union[trt.EngineCapability, EngineCapability] ) -> Optional[EngineCapability]: """Create a Torch-TensorRT engine capability enum from a TensorRT engine capability enum. @@ -1702,7 +1720,7 @@

    Source code for torch_tensorrt._enums

                 )
                 return None
    -
    [docs] def to( +
    [docs] def to( self, t: Union[Type[trt.EngineCapability], Type[EngineCapability]] ) -> Union[trt.EngineCapability, EngineCapability]: """Convert ``EngineCapability`` into the equivalent type in tensorrt @@ -1743,7 +1761,7 @@

    Source code for torch_tensorrt._enums

                 return self
     
             elif ENABLED_FEATURES.torchscript_frontend:
    -            from torch_tensorrt import _C
    +            from torch_tensorrt import _C
     
                 if t == _C.EngineCapability:
                     if self == EngineCapability.STANDARD:
    @@ -1759,7 +1777,7 @@ 

    Source code for torch_tensorrt._enums

                 "Provided unsupported destination type for engine capability type conversion"
             )
    -
    [docs] def try_to( +
    [docs] def try_to( self, t: Union[Type[trt.EngineCapability], Type[EngineCapability]] ) -> Optional[Union[trt.EngineCapability, EngineCapability]]: """Convert ``EngineCapability`` into the equivalent type in tensorrt @@ -1790,15 +1808,15 @@

    Source code for torch_tensorrt._enums

                 )
                 return None
    - def __eq__(self, other: Union[trt.EngineCapability, EngineCapability]) -> bool: + def __eq__(self, other: Union[trt.EngineCapability, EngineCapability]) -> bool: other_ = EngineCapability._from(other) return bool(self.value == other_.value) - def __hash__(self) -> int: + def __hash__(self) -> int: return hash(self.value)
    -class Platform(Enum): +class Platform(Enum): """ Specifies a target OS and CPU architecture that a Torch-TensorRT program targets """ @@ -1827,14 +1845,14 @@

    Source code for torch_tensorrt._enums

         UNKNOWN = auto()
     
         @classmethod
    -    def current_platform(cls) -> Platform:
    +    def current_platform(cls) -> Platform:
             """
             Returns an enum for the current platform Torch-TensorRT is running on
     
             Returns:
                 Platform: Current platform
             """
    -        import platform
    +        import platform
     
             if platform.system().lower().startswith("linux"):
                 # linux
    @@ -1850,11 +1868,11 @@ 

    Source code for torch_tensorrt._enums

     
             return Platform.UNKNOWN
     
    -    def __str__(self) -> str:
    +    def __str__(self) -> str:
             return str(self.name)
     
         @needs_torch_tensorrt_runtime  # type: ignore
    -    def _to_serialized_rt_platform(self) -> str:
    +    def _to_serialized_rt_platform(self) -> str:
             val: str = torch.ops.tensorrt._platform_unknown()
     
             if self == Platform.LINUX_X86_64:
    @@ -2091,7 +2109,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -2107,6 +2125,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -2145,6 +2166,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -2158,6 +2182,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_modules/torch_tensorrt/dynamo/_compiler.html b/docs/_modules/torch_tensorrt/dynamo/_compiler.html index 707265676e..fe0a41a04d 100644 --- a/docs/_modules/torch_tensorrt/dynamo/_compiler.html +++ b/docs/_modules/torch_tensorrt/dynamo/_compiler.html @@ -9,7 +9,7 @@ - torch_tensorrt.dynamo._compiler — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt.dynamo._compiler — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -132,8 +132,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -154,6 +154,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -201,6 +204,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -216,7 +223,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -272,7 +290,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -467,45 +485,45 @@

    Source code for torch_tensorrt.dynamo._compiler

    -from __future__ import annotations
    -
    -import collections.abc
    -import logging
    -import platform
    -import warnings
    -from typing import Any, Collection, List, Optional, Sequence, Set, Tuple, Union
    -
    -import torch
    -from torch.export import ExportedProgram
    -from torch.fx.node import Target
    -from torch_tensorrt._Device import Device
    -from torch_tensorrt._enums import EngineCapability, dtype
    -from torch_tensorrt._Input import Input
    -from torch_tensorrt.dynamo import _defaults, partitioning
    -from torch_tensorrt.dynamo._DryRunTracker import (
    +from __future__ import annotations
    +
    +import collections.abc
    +import logging
    +import platform
    +import warnings
    +from typing import Any, Collection, List, Optional, Sequence, Set, Tuple, Union
    +
    +import torch
    +from torch.export import ExportedProgram
    +from torch.fx.node import Target
    +from torch_tensorrt._Device import Device
    +from torch_tensorrt._enums import EngineCapability, dtype
    +from torch_tensorrt._Input import Input
    +from torch_tensorrt.dynamo import _defaults, partitioning
    +from torch_tensorrt.dynamo._DryRunTracker import (
         DryRunTracker,
         PerSubgraphData,
         dryrun_stats_display,
         parse_non_trt_nodes,
     )
    -from torch_tensorrt.dynamo._engine_cache import BaseEngineCache, DiskEngineCache
    -from torch_tensorrt.dynamo._exporter import replace_execute_engine_no_op_node
    -from torch_tensorrt.dynamo.conversion import (
    +from torch_tensorrt.dynamo._engine_cache import BaseEngineCache, DiskEngineCache
    +from torch_tensorrt.dynamo._exporter import replace_execute_engine_no_op_node
    +from torch_tensorrt.dynamo.conversion import (
         CompilationSettings,
         UnsupportedOperatorException,
         convert_module,
         interpret_module_to_result,
         repair_double_inputs,
     )
    -from torch_tensorrt.dynamo.conversion._ConverterRegistry import (
    +from torch_tensorrt.dynamo.conversion._ConverterRegistry import (
         DYNAMO_CONVERTERS as CONVERTERS,
     )
    -from torch_tensorrt.dynamo.lowering import (
    +from torch_tensorrt.dynamo.lowering import (
         get_decompositions,
         post_lowering,
         pre_export_lowering,
     )
    -from torch_tensorrt.dynamo.utils import (
    +from torch_tensorrt.dynamo.utils import (
         get_flat_args_with_check,
         get_output_metadata,
         parse_graph_io,
    @@ -518,7 +536,7 @@ 

    Source code for torch_tensorrt.dynamo._compiler

    < logger = logging.getLogger(__name__) -def cross_compile_for_windows( +def cross_compile_for_windows( exported_program: ExportedProgram, inputs: Optional[Sequence[Sequence[Any]]] = None, *, @@ -835,7 +853,7 @@

    Source code for torch_tensorrt.dynamo._compiler

    < return trt_gm -
    [docs]def compile( +
    [docs]def compile( exported_program: ExportedProgram, inputs: Optional[Sequence[Sequence[Any]]] = None, *, @@ -1153,7 +1171,7 @@

    Source code for torch_tensorrt.dynamo._compiler

    < return trt_gm
    -def compile_module( +def compile_module( gm: torch.fx.GraphModule, sample_arg_inputs: Sequence[Input], sample_kwarg_inputs: Optional[dict[Any, Any]] = None, @@ -1210,7 +1228,7 @@

    Source code for torch_tensorrt.dynamo._compiler

    < f"Detected support for {num_supported_ops} operators out of {total_ops} in subgraph." ) - def contains_metadata(gm: torch.fx.GraphModule) -> bool: + def contains_metadata(gm: torch.fx.GraphModule) -> bool: for node in gm.graph.nodes: if node.op != "output" and (not node.meta) and "val" not in node.meta: logger.warning( @@ -1379,7 +1397,7 @@

    Source code for torch_tensorrt.dynamo._compiler

    < return partitioned_module -def convert_exported_program_to_serialized_trt_engine( +def convert_exported_program_to_serialized_trt_engine( exported_program: ExportedProgram, inputs: Optional[Sequence[Sequence[Any]]] = None, *, @@ -1641,7 +1659,7 @@

    Source code for torch_tensorrt.dynamo._compiler

    < return serialized_engine -def save_cross_compiled_exported_program( +def save_cross_compiled_exported_program( gm: torch.fx.GraphModule, file_path: str, ) -> None: @@ -1655,14 +1673,14 @@

    Source code for torch_tensorrt.dynamo._compiler

    < if not file_path: raise ValueError("File path cannot be empty. Please provide a valid file path") - from torch_tensorrt.dynamo._exporter import export + from torch_tensorrt.dynamo._exporter import export exp_program = export(gm, cross_compile_flag=True) torch.export.save(exp_program, file_path) logger.debug(f"successfully saved the module for windows at {file_path}") -def load_cross_compiled_exported_program(file_path: str = "") -> Any: +def load_cross_compiled_exported_program(file_path: str = "") -> Any: """ Load an ExportedProgram file in Windows which was previously cross compiled in Linux @@ -1920,7 +1938,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1936,6 +1954,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1974,6 +1995,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1987,6 +2011,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_modules/torch_tensorrt/dynamo/_exporter.html b/docs/_modules/torch_tensorrt/dynamo/_exporter.html index cef5482e0c..334d362c19 100644 --- a/docs/_modules/torch_tensorrt/dynamo/_exporter.html +++ b/docs/_modules/torch_tensorrt/dynamo/_exporter.html @@ -9,7 +9,7 @@ - torch_tensorrt.dynamo._exporter — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt.dynamo._exporter — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -132,8 +132,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -154,6 +154,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -201,6 +204,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -216,7 +223,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -272,7 +290,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -467,16 +485,16 @@

    Source code for torch_tensorrt.dynamo._exporter

    -import base64
    -import copy
    -import operator
    -from typing import Any, Dict, Optional, Sequence, Tuple, cast
    -
    -import torch
    -from torch._guards import detect_fake_mode
    -from torch._subclasses.fake_tensor import FakeTensor
    -from torch.export import ExportedProgram, ExportGraphSignature
    -from torch.export.exported_program import (
    +import base64
    +import copy
    +import operator
    +from typing import Any, Dict, Optional, Sequence, Tuple, cast
    +
    +import torch
    +from torch._guards import detect_fake_mode
    +from torch._subclasses.fake_tensor import FakeTensor
    +from torch.export import ExportedProgram, ExportGraphSignature
    +from torch.export.exported_program import (
         CustomObjArgument,
         InputKind,
         InputSpec,
    @@ -486,10 +504,10 @@ 

    Source code for torch_tensorrt.dynamo._exporter

    < OutputSpec, TensorArgument, ) -from torch_tensorrt.dynamo.runtime._TorchTensorRTModule import ENGINE_IDX, NAME_IDX +from torch_tensorrt.dynamo.runtime._TorchTensorRTModule import ENGINE_IDX, NAME_IDX -
    [docs]def export( +
    [docs]def export( gm: torch.fx.GraphModule, cross_compile_flag: Optional[bool] = False, ) -> ExportedProgram: @@ -505,7 +523,7 @@

    Source code for torch_tensorrt.dynamo._exporter

    < return exp_program
    -def transform( +def transform( gm: torch.fx.GraphModule, cross_compile_flag: Optional[bool] = False, ) -> torch.fx.GraphModule: @@ -539,7 +557,7 @@

    Source code for torch_tensorrt.dynamo._exporter

    < return gm -def lift( +def lift( gm: torch.fx.GraphModule, graph_signature: Any ) -> Tuple[torch.fx.GraphModule, ExportGraphSignature, Dict[str, Any], Dict[str, Any]]: """ @@ -661,7 +679,7 @@

    Source code for torch_tensorrt.dynamo._exporter

    < return gm, graph_signature, state_dict, constants -def get_duplicate_nodes( +def get_duplicate_nodes( gm: torch.fx.GraphModule, submodule: torch.fx.GraphModule ) -> Tuple[Sequence[Any], Sequence[Any]]: """ @@ -684,7 +702,7 @@

    Source code for torch_tensorrt.dynamo._exporter

    < return submodule_duplicate_inputs, gm_duplicate_inputs -def inline_torch_modules(gm: torch.fx.GraphModule) -> torch.fx.GraphModule: +def inline_torch_modules(gm: torch.fx.GraphModule) -> torch.fx.GraphModule: """ Inline a submodule within the parent graph (gm). All `call_module` nodes should be replaced by their nodes in the submodule. @@ -751,7 +769,7 @@

    Source code for torch_tensorrt.dynamo._exporter

    < return gm -def copy_submodule_attributes( +def copy_submodule_attributes( gm: torch.fx.GraphModule, submodule: torch.fx.GraphModule, submodule_name: str ) -> None: """ @@ -762,7 +780,7 @@

    Source code for torch_tensorrt.dynamo._exporter

    < _assign_attr does exactly that. It creates a module for eg: conv, adds an attribute weight to it and adds this conv module as an attribute to parent gm. """ - from torch.export.unflatten import _assign_attr, _AttrKind + from torch.export.unflatten import _assign_attr, _AttrKind for key, value in submodule.named_parameters(): _assign_attr(value, gm, key, _AttrKind.PARAMETER) @@ -771,7 +789,7 @@

    Source code for torch_tensorrt.dynamo._exporter

    < _assign_attr(value, gm, key, _AttrKind.BUFFER) -def create_trt_exp_program( +def create_trt_exp_program( gm: torch.fx.GraphModule, ) -> ExportedProgram: """Creates a new Exported Program. This function takes an torch.fx.GraphModule which has TRT engines @@ -825,7 +843,7 @@

    Source code for torch_tensorrt.dynamo._exporter

    < return trt_exp_program -def inline_trt_modules( +def inline_trt_modules( gm: torch.fx.GraphModule, cross_compile_flag: Optional[bool] = False ) -> torch.fx.GraphModule: """ @@ -901,7 +919,7 @@

    Source code for torch_tensorrt.dynamo._exporter

    < return gm -def replace_execute_engine_no_op_node( +def replace_execute_engine_no_op_node( exp_program: ExportedProgram, ) -> ExportedProgram: gm = exp_program.graph_module @@ -1179,7 +1197,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1195,6 +1213,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1233,6 +1254,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1246,6 +1270,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_modules/torch_tensorrt/dynamo/_refit.html b/docs/_modules/torch_tensorrt/dynamo/_refit.html index 48beb9d4bf..0bafebf55b 100644 --- a/docs/_modules/torch_tensorrt/dynamo/_refit.html +++ b/docs/_modules/torch_tensorrt/dynamo/_refit.html @@ -9,7 +9,7 @@ - torch_tensorrt.dynamo._refit — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt.dynamo._refit — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -132,8 +132,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -154,6 +154,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -201,6 +204,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -216,7 +223,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -272,7 +290,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -467,42 +485,42 @@

    Source code for torch_tensorrt.dynamo._refit

    -from __future__ import annotations
    -
    -import collections.abc
    -import copy
    -import logging
    -from typing import Any, List, Optional, Sequence, Tuple
    -
    -import numpy as np
    -import tensorrt as trt
    -import torch
    -from torch.export import ExportedProgram
    -from torch_tensorrt._enums import dtype
    -from torch_tensorrt._Input import Input
    -from torch_tensorrt.dynamo import partitioning
    -from torch_tensorrt.dynamo._exporter import inline_torch_modules
    -from torch_tensorrt.dynamo._settings import CompilationSettings
    -from torch_tensorrt.dynamo.conversion._conversion import infer_module_output_dtypes
    -from torch_tensorrt.dynamo.conversion._ConverterRegistry import (
    +from __future__ import annotations
    +
    +import collections.abc
    +import copy
    +import logging
    +from typing import Any, List, Optional, Sequence, Tuple
    +
    +import numpy as np
    +import tensorrt as trt
    +import torch
    +from torch.export import ExportedProgram
    +from torch_tensorrt._enums import dtype
    +from torch_tensorrt._Input import Input
    +from torch_tensorrt.dynamo import partitioning
    +from torch_tensorrt.dynamo._exporter import inline_torch_modules
    +from torch_tensorrt.dynamo._settings import CompilationSettings
    +from torch_tensorrt.dynamo.conversion._conversion import infer_module_output_dtypes
    +from torch_tensorrt.dynamo.conversion._ConverterRegistry import (
         DYNAMO_CONVERTERS as CONVERTERS,
     )
    -from torch_tensorrt.dynamo.conversion._TRTInterpreter import TRTInterpreter
    -from torch_tensorrt.dynamo.conversion.truncate_double import repair_double_inputs
    -from torch_tensorrt.dynamo.lowering import (
    +from torch_tensorrt.dynamo.conversion._TRTInterpreter import TRTInterpreter
    +from torch_tensorrt.dynamo.conversion.truncate_double import repair_double_inputs
    +from torch_tensorrt.dynamo.lowering import (
         get_decompositions,
         post_lowering,
         pre_export_lowering,
     )
    -from torch_tensorrt.dynamo.runtime._PythonTorchTensorRTModule import (
    +from torch_tensorrt.dynamo.runtime._PythonTorchTensorRTModule import (
         PythonTorchTensorRTModule,
     )
    -from torch_tensorrt.dynamo.runtime._TorchTensorRTModule import (
    +from torch_tensorrt.dynamo.runtime._TorchTensorRTModule import (
         ENGINE_IDX,
         SERIALIZED_METADATA_IDX,
         TorchTensorRTModule,
     )
    -from torch_tensorrt.dynamo.utils import (
    +from torch_tensorrt.dynamo.utils import (
         check_module_output,
         get_model_device,
         get_torch_inputs,
    @@ -510,12 +528,12 @@ 

    Source code for torch_tensorrt.dynamo._refit

    to_torch_device,
         to_torch_tensorrt_device,
     )
    -from torch_tensorrt.logging import TRT_LOGGER
    +from torch_tensorrt.logging import TRT_LOGGER
     
     logger = logging.getLogger(__name__)
     
     
    -def construct_refit_mapping(
    +def construct_refit_mapping(
         module: torch.fx.GraphModule,
         inputs: Sequence[Input],
         settings: CompilationSettings = CompilationSettings(),
    @@ -576,7 +594,7 @@ 

    Source code for torch_tensorrt.dynamo._refit

    return weight_map
     
     
    -def construct_refit_mapping_from_weight_name_map(
    +def construct_refit_mapping_from_weight_name_map(
         weight_name_map: dict[Any, Any], state_dict: dict[Any, Any]
     ) -> dict[Any, Any]:
         engine_weight_map = {}
    @@ -602,7 +620,7 @@ 

    Source code for torch_tensorrt.dynamo._refit

    return engine_weight_map
     
     
    -def _refit_single_trt_engine_with_gm(
    +def _refit_single_trt_engine_with_gm(
         new_gm: torch.fx.GraphModule,
         old_engine: trt.ICudaEngine,
         input_list: Sequence[Any],
    @@ -680,7 +698,7 @@ 

    Source code for torch_tensorrt.dynamo._refit

    raise AssertionError("Refitting failed.")
     
     
    -
    [docs]def refit_module_weights( +
    [docs]def refit_module_weights( compiled_module: torch.fx.GraphModule | ExportedProgram, new_weight_module: ExportedProgram, arg_inputs: Optional[Tuple[Any, ...]] = None, @@ -979,10 +997,10 @@

    Source code for torch_tensorrt.dynamo._refit

    # Util functions -----------
    -import base64
    +import base64
     
     
    -def get_engine_from_encoded_engine(
    +def get_engine_from_encoded_engine(
         encoded_engine: str, runtime: trt.Runtime
     ) -> trt.ICudaEngine:
         serialized_engine = base64.b64decode(encoded_engine)
    @@ -1214,7 +1232,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1230,6 +1248,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1268,6 +1289,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1281,6 +1305,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_modules/torch_tensorrt/dynamo/_settings.html b/docs/_modules/torch_tensorrt/dynamo/_settings.html index a1e632dce0..6448e409f2 100644 --- a/docs/_modules/torch_tensorrt/dynamo/_settings.html +++ b/docs/_modules/torch_tensorrt/dynamo/_settings.html @@ -9,7 +9,7 @@ - torch_tensorrt.dynamo._settings — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt.dynamo._settings — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -132,8 +132,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -154,6 +154,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -201,6 +204,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -216,7 +223,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -272,7 +290,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -467,13 +485,13 @@

    Source code for torch_tensorrt.dynamo._settings

    -from dataclasses import dataclass, field
    -from typing import Collection, Optional, Set, Tuple, Union
    +from dataclasses import dataclass, field
    +from typing import Collection, Optional, Set, Tuple, Union
     
    -from torch.fx.node import Target
    -from torch_tensorrt._Device import Device
    -from torch_tensorrt._enums import EngineCapability, dtype
    -from torch_tensorrt.dynamo._defaults import (
    +from torch.fx.node import Target
    +from torch_tensorrt._Device import Device
    +from torch_tensorrt._enums import EngineCapability, dtype
    +from torch_tensorrt.dynamo._defaults import (
         ASSUME_DYNAMIC_SHAPE_SUPPORT,
         CACHE_BUILT_ENGINES,
         DEBUG,
    @@ -502,6 +520,7 @@ 

    Source code for torch_tensorrt.dynamo._settings

    < STRIP_ENGINE_WEIGHTS, TIMING_CACHE_PATH, TRUNCATE_DOUBLE, + USE_AOT_JOINT_EXPORT, USE_EXPLICIT_TYPING, USE_FAST_PARTITIONER, USE_FP32_ACC, @@ -513,7 +532,7 @@

    Source code for torch_tensorrt.dynamo._settings

    <
    [docs]@dataclass -class CompilationSettings: +class CompilationSettings: """Compilation settings for Torch-TensorRT Dynamo Paths Args: @@ -560,6 +579,7 @@

    Source code for torch_tensorrt.dynamo._settings

    < enable_weight_streaming (bool): Enable weight streaming. enable_cross_compile_for_windows (bool): By default this is False means TensorRT engines can only be executed on the same platform where they were built. True will enable cross-platform compatibility which allows the engine to be built on Linux and run on Windows + use_aot_joint_export (bool): Use aot_export_joint_simple, else wrap backend with AOT_autograd, required for distributed tensors """ enabled_precisions: Set[dtype] = field(default_factory=lambda: ENABLED_PRECISIONS) @@ -599,7 +619,8 @@

    Source code for torch_tensorrt.dynamo._settings

    < strip_engine_weights: bool = STRIP_ENGINE_WEIGHTS immutable_weights: bool = IMMUTABLE_WEIGHTS enable_weight_streaming: bool = ENABLE_WEIGHT_STREAMING - enable_cross_compile_for_windows: bool = ENABLE_CROSS_COMPILE_FOR_WINDOWS
    + enable_cross_compile_for_windows: bool = ENABLE_CROSS_COMPILE_FOR_WINDOWS + use_aot_joint_export: bool = USE_AOT_JOINT_EXPORT
    _SETTINGS_TO_BE_ENGINE_INVARIANT = ( @@ -618,7 +639,7 @@

    Source code for torch_tensorrt.dynamo._settings

    < ) -def settings_are_compatible( +def settings_are_compatible( set_a: CompilationSettings, set_b: CompilationSettings ) -> Tuple[bool, Set[str]]: incompatible_settings: Set[str] = set() @@ -857,7 +878,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -873,6 +894,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -911,6 +935,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -924,6 +951,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_modules/torch_tensorrt/dynamo/_tracer.html b/docs/_modules/torch_tensorrt/dynamo/_tracer.html index 469fc51fb0..c1324f6719 100644 --- a/docs/_modules/torch_tensorrt/dynamo/_tracer.html +++ b/docs/_modules/torch_tensorrt/dynamo/_tracer.html @@ -9,7 +9,7 @@ - torch_tensorrt.dynamo._tracer — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt.dynamo._tracer — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -132,8 +132,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -154,6 +154,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -201,6 +204,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -216,7 +223,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -272,7 +290,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -467,22 +485,22 @@

    Source code for torch_tensorrt.dynamo._tracer

    -from __future__ import annotations
    +from __future__ import annotations
     
    -import logging
    -from inspect import signature
    -from typing import Any, Optional, Tuple, Union
    +import logging
    +from inspect import signature
    +from typing import Any, Optional, Tuple, Union
     
    -import torch
    -from torch.export import Dim, export
    -from torch_tensorrt._Input import Input
    -from torch_tensorrt.dynamo._defaults import DEBUG, default_device
    -from torch_tensorrt.dynamo.utils import get_torch_inputs, set_log_level, to_torch_device
    +import torch
    +from torch.export import Dim, export
    +from torch_tensorrt._Input import Input
    +from torch_tensorrt.dynamo._defaults import DEBUG, default_device
    +from torch_tensorrt.dynamo.utils import get_torch_inputs, set_log_level, to_torch_device
     
     logger = logging.getLogger(__name__)
     
     
    -
    [docs]def trace( +
    [docs]def trace( mod: torch.nn.Module | torch.fx.GraphModule, inputs: Optional[Tuple[Any, ...]] = None, *, @@ -559,7 +577,7 @@

    Source code for torch_tensorrt.dynamo._tracer

    return exp_program
    -def get_dynamic_shapes_kwargs(inputs: Any) -> Union[dict[str, Any], list[Any]]: +def get_dynamic_shapes_kwargs(inputs: Any) -> Union[dict[str, Any], list[Any]]: if isinstance(inputs, dict): dynamic_shapes_kwarg = {} for k, v in inputs.items(): @@ -578,7 +596,7 @@

    Source code for torch_tensorrt.dynamo._tracer

    raise TypeError(f"Unknown type {type(inputs)}.") -def get_dynamic_shapes_args(mod: torch.nn.Module, inputs: Any) -> dict[str, Any]: +def get_dynamic_shapes_args(mod: torch.nn.Module, inputs: Any) -> dict[str, Any]: # dynamic_shape is a dict and cannot work without keys. Here we use position argument name # in forward function as the name args = list(signature(mod.forward).parameters.keys()) @@ -588,7 +606,7 @@

    Source code for torch_tensorrt.dynamo._tracer

    return dynamic_shapes -def get_dynamic_shapes(input: Input) -> dict[Any, Any]: +def get_dynamic_shapes(input: Input) -> dict[Any, Any]: if not isinstance(input, Input): # If the input is torch.Tensor, no dynamic is needed. Return empty dict return {} @@ -835,7 +853,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -851,6 +869,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -889,6 +910,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -902,6 +926,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_modules/torch_tensorrt/dynamo/runtime/_MutableTorchTensorRTModule.html b/docs/_modules/torch_tensorrt/dynamo/runtime/_MutableTorchTensorRTModule.html index e1684e2e14..5a8be28457 100644 --- a/docs/_modules/torch_tensorrt/dynamo/runtime/_MutableTorchTensorRTModule.html +++ b/docs/_modules/torch_tensorrt/dynamo/runtime/_MutableTorchTensorRTModule.html @@ -9,7 +9,7 @@ - torch_tensorrt.dynamo.runtime._MutableTorchTensorRTModule — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt.dynamo.runtime._MutableTorchTensorRTModule — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -132,8 +132,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -154,6 +154,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -201,6 +204,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -216,7 +223,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -272,7 +290,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -467,21 +485,21 @@

    Source code for torch_tensorrt.dynamo.runtime._MutableTorchTensorRTModule

    -import logging
    -from copy import deepcopy
    -from enum import Enum, auto
    -from typing import Any, Collection, Dict, Iterator, List, Optional, Set, Union
    -
    -import numpy as np
    -import torch
    -from torch.fx.node import Target
    -from torch_tensorrt._Device import Device
    -from torch_tensorrt._enums import EngineCapability, dtype
    -from torch_tensorrt.dynamo import _defaults
    -from torch_tensorrt.dynamo._compiler import compile as dynamo_compile
    -from torch_tensorrt.dynamo._refit import refit_module_weights
    -from torch_tensorrt.dynamo._settings import CompilationSettings
    -from torch_tensorrt.dynamo.utils import (
    +import logging
    +from copy import deepcopy
    +from enum import Enum, auto
    +from typing import Any, Collection, Dict, Iterator, List, Optional, Set, Union
    +
    +import numpy as np
    +import torch
    +from torch.fx.node import Target
    +from torch_tensorrt._Device import Device
    +from torch_tensorrt._enums import EngineCapability, dtype
    +from torch_tensorrt.dynamo import _defaults
    +from torch_tensorrt.dynamo._compiler import compile as dynamo_compile
    +from torch_tensorrt.dynamo._refit import refit_module_weights
    +from torch_tensorrt.dynamo._settings import CompilationSettings
    +from torch_tensorrt.dynamo.utils import (
         check_output_equal,
         to_torch_device,
         to_torch_tensorrt_device,
    @@ -490,27 +508,27 @@ 

    Source code for torch_tensorrt.dynamo.runtime._MutableTorchTensorRTModulelogger = logging.getLogger(__name__) -class RefitFlag(Enum): +class RefitFlag(Enum): UNKNOWN = auto() NEEDS_REFIT = auto() NEEDS_RECOMPILE = auto() LIVE = auto() -class RefitState: +class RefitState: _state: RefitFlag = RefitFlag.NEEDS_RECOMPILE - def set_state(self, state: RefitFlag) -> None: + def set_state(self, state: RefitFlag) -> None: if isinstance(state, RefitFlag): self._state = state else: raise ValueError(f"Invalid state: {state}") - def get_state(self) -> RefitFlag: + def get_state(self) -> RefitFlag: return self._state -
    [docs]class MutableTorchTensorRTModule(object): +
    [docs]class MutableTorchTensorRTModule(object): """ Initialize a MutableTorchTensorRTModule to seamlessly manipulate it like a regular PyTorch module. All TensorRT compilation and refitting processes are handled automatically as you work with the module. @@ -522,7 +540,7 @@

    Source code for torch_tensorrt.dynamo.runtime._MutableTorchTensorRTModule Any modifications made to the MutableTorchTensorRTModule will be reflected in both the TensorRT graph module and the original PyTorch module. """ -
    [docs] def __init__( +
    [docs] def __init__( self, pytorch_model: torch.nn.Module, *, @@ -672,22 +690,22 @@

    Source code for torch_tensorrt.dynamo.runtime._MutableTorchTensorRTModule) self.init_finished = True

    - def store_state_dict_metadata(self) -> None: + def store_state_dict_metadata(self) -> None: for k, v in self.original_model.state_dict().items(): self.state_dict_metadata[k] = v.shape - def load_state_dict( + def load_state_dict( self, state_dict: Dict[str, Any], strict: bool = True, assign: bool = False ) -> None: self.refit_state.set_state(RefitFlag.NEEDS_REFIT) self.original_model.load_state_dict(state_dict, strict=strict, assign=assign) @staticmethod - def _transform_state_dict(sd: Dict[str, Any]) -> Dict[str, torch.nn.Parameter]: + def _transform_state_dict(sd: Dict[str, Any]) -> Dict[str, torch.nn.Parameter]: return {k: torch.nn.Parameter(v, requires_grad=False) for k, v in sd.items()} - def update_refit_condition(self) -> None: + def update_refit_condition(self) -> None: # 2-stage check to determine whether the module should be intact, refitted, or recompiled. # Default refit @@ -721,7 +739,7 @@

    Source code for torch_tensorrt.dynamo.runtime._MutableTorchTensorRTModulereturn -
    [docs] def refit_gm(self) -> None: +
    [docs] def refit_gm(self) -> None: """ Refit the TRT graph module with any updates. This function should be called whenever the weight values get changed but the weight structure remains @@ -752,7 +770,7 @@

    Source code for torch_tensorrt.dynamo.runtime._MutableTorchTensorRTModuleself.original_model.cpu() torch.cuda.empty_cache()

    -
    [docs] def compile(self) -> None: +
    [docs] def compile(self) -> None: """ (Re)compile the TRT graph module using the PyTorch module. This function should be called whenever the weight structure get changed (shape, more layers...) @@ -775,7 +793,7 @@

    Source code for torch_tensorrt.dynamo.runtime._MutableTorchTensorRTModuleself.original_model.cpu() torch.cuda.empty_cache()

    - def _validate_inputs(self, *args: Any, **kwargs: Any) -> None: + def _validate_inputs(self, *args: Any, **kwargs: Any) -> None: if ( not self.arg_inputs or not MutableTorchTensorRTModule.check_inputs_equal(self.arg_inputs, args) @@ -787,12 +805,12 @@

    Source code for torch_tensorrt.dynamo.runtime._MutableTorchTensorRTModuleself.refit_state.set_state(RefitFlag.NEEDS_RECOMPILE) self.store_inputs(args, kwargs) - def store_inputs(self, arg_inputs: Any, kwarg_inputs: Any) -> None: + def store_inputs(self, arg_inputs: Any, kwarg_inputs: Any) -> None: self.arg_inputs = arg_inputs self.kwarg_inputs = kwarg_inputs @staticmethod - def process_kwarg_inputs(inputs: Any) -> Any: + def process_kwarg_inputs(inputs: Any) -> Any: # Process kwarg inputs to be acceptable for Torch-TensorRT if isinstance(inputs, dict): # None should be excluded. AOT compile also does not allow dynamic control flow, bool is also excluded. @@ -816,7 +834,7 @@

    Source code for torch_tensorrt.dynamo.runtime._MutableTorchTensorRTModule+ "Allowed input types: {torch_tensorrt.Input, torch.Tensor, list, tuple, dict}" ) - def forward(self, *args: Any, **kwargs: Any) -> Any: + def forward(self, *args: Any, **kwargs: Any) -> Any: # Step 1: Check whether the input shape has changed kwargs = MutableTorchTensorRTModule.process_kwarg_inputs(kwargs) self._validate_inputs(*args, **kwargs) @@ -849,11 +867,11 @@

    Source code for torch_tensorrt.dynamo.runtime._MutableTorchTensorRTModuleself.run_info = (args, kwargs, result) return result - def to(self, device: str) -> None: + def to(self, device: str) -> None: logger.warning("Original PyTorch model is moved. CPU offload may failed.") self.orignial_model.to(device) - def __deepcopy__(self, memo: Any) -> Any: + def __deepcopy__(self, memo: Any) -> Any: cls = self.__class__ result = cls.__new__(cls) memo[id(self)] = result @@ -865,10 +883,10 @@

    Source code for torch_tensorrt.dynamo.runtime._MutableTorchTensorRTModule) return result - def __call__(self, *args: Any, **kwargs: Any) -> Any: + def __call__(self, *args: Any, **kwargs: Any) -> Any: return self.forward(*args, **kwargs) - def __getattr__(self, name: str) -> Any: + def __getattr__(self, name: str) -> Any: if name in self.__dict__: # this object has it @@ -881,7 +899,7 @@

    Source code for torch_tensorrt.dynamo.runtime._MutableTorchTensorRTModulereturn getattr(self.pytorch_model, name) - def __delattr__(self, name: str) -> Any: + def __delattr__(self, name: str) -> Any: if name in self.__dict__: # this object has it @@ -889,7 +907,7 @@

    Source code for torch_tensorrt.dynamo.runtime._MutableTorchTensorRTModulereturn self.pytorch_model.__delattr__(name) - def __setattr__(self, name: str, value: Any) -> None: + def __setattr__(self, name: str, value: Any) -> None: # When the module finished initialization, any modification to attributes that does not exist # in __dict__ will be handled in pytorch module. if self.init_finished: @@ -905,7 +923,7 @@

    Source code for torch_tensorrt.dynamo.runtime._MutableTorchTensorRTModuleobject.__setattr__(self, name, value) @staticmethod - def check_inputs_equal( + def check_inputs_equal( input1: Any, input2: Any, ) -> bool: @@ -938,7 +956,7 @@

    Source code for torch_tensorrt.dynamo.runtime._MutableTorchTensorRTModulereturn True @staticmethod - def save(module: Any, path: str) -> None: + def save(module: Any, path: str) -> None: # Cast the object back to MutableTorchTensorRTModule to save assert ( not module.settings.use_python_runtime @@ -964,7 +982,7 @@

    Source code for torch_tensorrt.dynamo.runtime._MutableTorchTensorRTModulemodule.init_finished = True @staticmethod - def load(path: str) -> Any: + def load(path: str) -> Any: # When the model get saved, init_finished is set to False. # Class is restored to MutableTorchTensorRTModule, and some attribute is deleted module = torch.load(path, weights_only=False) @@ -986,7 +1004,7 @@

    Source code for torch_tensorrt.dynamo.runtime._MutableTorchTensorRTModulereturn module

    -def recursively_remove_trigger(obj: Any) -> Any: +def recursively_remove_trigger(obj: Any) -> Any: # Not safe: If the object has a circular reference (such as a doubly linkded list), this will cause infinite recursion if obj.__class__.__name__ == "ChangeTriggerWrapper": obj = obj.instance @@ -1008,18 +1026,18 @@

    Source code for torch_tensorrt.dynamo.runtime._MutableTorchTensorRTModulereturn obj -def _make_refit_change_trigger(obj: object, refit_state: RefitState) -> Any: +def _make_refit_change_trigger(obj: object, refit_state: RefitState) -> Any: subclass: type = obj.__class__ - class ChangeTriggerWrapper(subclass): # type: ignore + class ChangeTriggerWrapper(subclass): # type: ignore # The reason why we want to inherent to the subclass is that we want the ChangeTriggerWrapper shares all functions # that an ordinary object has. In this way attributes accessed inside a function will be from the __getattr__function # of ChangeTriggerWrapper, instead of the object itself, thus be recursively wrapped by ChangeTriggerWrapper. - def __init__(self, obj: Any): + def __init__(self, obj: Any): object.__setattr__(self, "instance", obj) - def __getattr__( + def __getattr__( self, name: str ) -> Any: # Called when the attribute does not exist obj = getattr(self.instance, name) @@ -1034,7 +1052,7 @@

    Source code for torch_tensorrt.dynamo.runtime._MutableTorchTensorRTModulereturn _make_refit_change_trigger(obj, refit_state) return obj - def __setattr__(self, name: str, value: Any) -> None: + def __setattr__(self, name: str, value: Any) -> None: # If we need to set __dict__ or instance, we directly set it to the trigger wrapper. # Enable setting __dict__ is because PyTorch proxy uses __new__ to initialize a shallow copy # of a module and explicit set the __dict__. If we don't set __dict__ it will get infinite recursion. @@ -1046,44 +1064,44 @@

    Source code for torch_tensorrt.dynamo.runtime._MutableTorchTensorRTModulevalue = recursively_remove_trigger(value) setattr(self.instance, name, value) - def __delattr__(self, name: str) -> None: + def __delattr__(self, name: str) -> None: self._on_change() delattr( self.instance, name, ) - def _on_change(self) -> None: + def _on_change(self) -> None: refit_state.set_state(RefitFlag.UNKNOWN) logger.info( "Attribute modification detected. The module will be refitted later." ) - def __call__(self, *args: Any, **kwargs: Any) -> Any: + def __call__(self, *args: Any, **kwargs: Any) -> Any: return self.instance(*args, **kwargs) - def _call_impl(self, *args: Any, **kwargs: Any) -> Any: + def _call_impl(self, *args: Any, **kwargs: Any) -> Any: return self.instance._call_impl(*args, **kwargs) - def forward(self, *args: Any, **kwargs: Any) -> Any: + def forward(self, *args: Any, **kwargs: Any) -> Any: return self.instance.forward(*args, **kwargs) - def __setitem__(self, item: str, value: Any) -> None: + def __setitem__(self, item: str, value: Any) -> None: self._on_change() # We want to make sure the original PyTorch model does not have a trigger wrapper value = recursively_remove_trigger(value) self.instance.__setitem__(item, value) - def __getitem__(self, items: str) -> Any: + def __getitem__(self, items: str) -> Any: obj = self.instance.__getitem__(items) if isinstance(obj, ChangeTriggerWrapper): return obj return _make_refit_change_trigger(obj, refit_state) - def __len__(self) -> int: + def __len__(self) -> int: return len(self.instance) - def __iter__(self) -> Iterator[Any]: + def __iter__(self) -> Iterator[Any]: return iter(self.instance) return ChangeTriggerWrapper(obj) @@ -1313,7 +1331,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1329,6 +1347,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1367,6 +1388,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1380,6 +1404,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_modules/torch_tensorrt/dynamo/runtime/_PythonTorchTensorRTModule.html b/docs/_modules/torch_tensorrt/dynamo/runtime/_PythonTorchTensorRTModule.html index 549d8fbe54..14d8aa834e 100644 --- a/docs/_modules/torch_tensorrt/dynamo/runtime/_PythonTorchTensorRTModule.html +++ b/docs/_modules/torch_tensorrt/dynamo/runtime/_PythonTorchTensorRTModule.html @@ -9,7 +9,7 @@ - torch_tensorrt.dynamo.runtime._PythonTorchTensorRTModule — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt.dynamo.runtime._PythonTorchTensorRTModule — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -132,8 +132,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -154,6 +154,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +

    @@ -201,6 +204,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +

    @@ -216,7 +223,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -272,7 +290,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -467,23 +485,23 @@

    Source code for torch_tensorrt.dynamo.runtime._PythonTorchTensorRTModule

    -from __future__ import annotations
    -
    -import logging
    -from contextlib import nullcontext
    -from tempfile import tempdir
    -from typing import Any, Dict, List, Optional, Sequence, Tuple
    -
    -import tensorrt as trt
    -import torch
    -import torch_tensorrt
    -from torch.nn import Module
    -from torch_tensorrt._Device import Device
    -from torch_tensorrt._enums import Platform, dtype
    -from torch_tensorrt.dynamo._settings import CompilationSettings
    -from torch_tensorrt.dynamo.utils import DYNAMIC_DIM
    -from torch_tensorrt.logging import TRT_LOGGER
    -from torch_tensorrt.runtime._utils import (
    +from __future__ import annotations
    +
    +import logging
    +from contextlib import nullcontext
    +from tempfile import tempdir
    +from typing import Any, Dict, List, Optional, Sequence, Tuple
    +
    +import tensorrt as trt
    +import torch
    +import torch_tensorrt
    +from torch.nn import Module
    +from torch_tensorrt._Device import Device
    +from torch_tensorrt._enums import Platform, dtype
    +from torch_tensorrt.dynamo._settings import CompilationSettings
    +from torch_tensorrt.dynamo.utils import DYNAMIC_DIM
    +from torch_tensorrt.logging import TRT_LOGGER
    +from torch_tensorrt.runtime._utils import (
         _is_switch_required,
         _select_rt_device,
         multi_gpu_device_check,
    @@ -492,8 +510,8 @@ 

    Source code for torch_tensorrt.dynamo.runtime._PythonTorchTensorRTModule

    logger = logging.getLogger(__name__) -class TorchTRTRuntimeStates: - def __init__(self, new_cudagraphs: bool): +class TorchTRTRuntimeStates: + def __init__(self, new_cudagraphs: bool): # Indicates whether CUDAGraphs were enabled in the previous execute_engine self.old_cudagraphs = new_cudagraphs # Indicates whether pre-allocated output was enabled in the previous execute_engine @@ -501,7 +519,7 @@

    Source code for torch_tensorrt.dynamo.runtime._PythonTorchTensorRTModule

    # Indicates whether context has changed self.context_changed = False - def set_runtime_states( + def set_runtime_states( self, new_cudagraphs: bool, new_pre_allocated_output: bool, @@ -545,14 +563,14 @@

    Source code for torch_tensorrt.dynamo.runtime._PythonTorchTensorRTModule

    ) -
    [docs]class PythonTorchTensorRTModule(Module): # type: ignore[misc] +
    [docs]class PythonTorchTensorRTModule(Module): # type: ignore[misc] """PythonTorchTensorRTModule is a PyTorch module which encompasses an arbitrary TensorRT Engine. This module is backed by the Torch-TensorRT runtime and is only compatible with FX / Dynamo / Python deployments. This module cannot be serialized to torchscript via torch.jit.trace for C++ deployment. """ -
    [docs] def __init__( +
    [docs] def __init__( self, serialized_engine: Optional[bytes] = None, input_binding_names: Optional[List[str]] = None, @@ -639,16 +657,16 @@

    Source code for torch_tensorrt.dynamo.runtime._PythonTorchTensorRTModule

    if self.serialized_engine is not None and not self.settings.lazy_engine_init: self.setup_engine()
    - def get_streamable_device_memory_budget(self) -> Any: + def get_streamable_device_memory_budget(self) -> Any: return self.engine.streamable_weights_size - def get_automatic_device_memory_budget(self) -> Any: + def get_automatic_device_memory_budget(self) -> Any: return self.engine.get_weight_streaming_automatic_budget() - def get_device_memory_budget(self) -> Any: + def get_device_memory_budget(self) -> Any: return self.engine.weight_streaming_budget_v2 - def set_device_memory_budget(self, budget_bytes: int) -> int: + def set_device_memory_budget(self, budget_bytes: int) -> int: # Recreating the context because weight streaming budget cannot be modified while there are active context. if self.context is not None: del self.context @@ -657,7 +675,7 @@

    Source code for torch_tensorrt.dynamo.runtime._PythonTorchTensorRTModule

    self.runtime_states.context_changed = True return budget_bytes - def _set_device_memory_budget(self, budget_bytes: int) -> int: + def _set_device_memory_budget(self, budget_bytes: int) -> int: # Disable weight streaming for invalid budget size if budget_bytes < 0: budget_bytes = self.get_streamable_device_memory_budget() @@ -670,13 +688,13 @@

    Source code for torch_tensorrt.dynamo.runtime._PythonTorchTensorRTModule

    return budget_bytes - def set_default_device_memory_budget(self) -> int: + def set_default_device_memory_budget(self) -> int: budget_bytes = self.get_automatic_device_memory_budget() # Set automatic weight streaming budget as default when context is created logger.debug(f"Weight streaming budget set to {budget_bytes}B") return self._set_device_memory_budget(budget_bytes) - def setup_engine(self) -> None: + def setup_engine(self) -> None: assert ( self.target_platform == Platform.current_platform() ), f"TensorRT engine was not built to target current platform (target: {self.target_platform}, current: {Platform.current_platform()})" @@ -710,17 +728,17 @@

    Source code for torch_tensorrt.dynamo.runtime._PythonTorchTensorRTModule

    if torch_tensorrt.runtime.get_cudagraphs_mode(): self.cudagraph = torch.cuda.CUDAGraph() - def _check_initialized(self) -> None: + def _check_initialized(self) -> None: if not self.initialized: raise RuntimeError("PythonTorchTensorRTModule is not initialized.") - def _on_state_dict(self, state_dict: Dict[str, Any], prefix: str, _: Any) -> None: + def _on_state_dict(self, state_dict: Dict[str, Any], prefix: str, _: Any) -> None: state_dict[prefix + "engine"] = self.serialized_engine state_dict[prefix + "input_names"] = self.input_names state_dict[prefix + "output_names"] = self.output_names state_dict[prefix + "platform"] = self.target_platform - def _load_from_state_dict( + def _load_from_state_dict( self, state_dict: Dict[str, Any], prefix: str, @@ -739,28 +757,28 @@

    Source code for torch_tensorrt.dynamo.runtime._PythonTorchTensorRTModule

    multi_gpu_device_check() self.setup_engine() - def __getstate__(self) -> Dict[str, Any]: + def __getstate__(self) -> Dict[str, Any]: state = self.__dict__.copy() state.pop("engine", None) state.pop("context", None) return state - def __setstate__(self, state: Dict[str, Any]) -> None: + def __setstate__(self, state: Dict[str, Any]) -> None: self.__dict__.update(state) self.setup_engine() - def __deepcopy__(self, memo: Any) -> PythonTorchTensorRTModule: + def __deepcopy__(self, memo: Any) -> PythonTorchTensorRTModule: cls = self.__class__ result = cls.__new__(cls) memo[id(self)] = result result.__setstate__(self.__getstate__()) return result - def __del__(self) -> None: + def __del__(self) -> None: if self.cudagraph: self.cudagraph.reset() - def setup_input_tensors( + def setup_input_tensors( self, contiguous_inputs: List[torch.Tensor], cudagraphs_enabled: bool, @@ -811,7 +829,7 @@

    Source code for torch_tensorrt.dynamo.runtime._PythonTorchTensorRTModule

    input_name, contiguous_inputs[i].data_ptr() ) - def create_output_tensors(self) -> List[torch.Tensor]: + def create_output_tensors(self) -> List[torch.Tensor]: # create output tensors outputs: List[torch.Tensor] = [] @@ -824,10 +842,10 @@

    Source code for torch_tensorrt.dynamo.runtime._PythonTorchTensorRTModule

    outputs.append(output) return outputs - def set_pre_allocated_outputs(self, enable: bool) -> None: + def set_pre_allocated_outputs(self, enable: bool) -> None: self.use_pre_allocated_outputs = enable -
    [docs] def forward(self, *inputs: torch.Tensor) -> torch.Tensor | Tuple[torch.Tensor, ...]: +
    [docs] def forward(self, *inputs: torch.Tensor) -> torch.Tensor | Tuple[torch.Tensor, ...]: # Ensure inputs are available in all scopes and cast symbolic integers to Tensors contiguous_inputs: List[torch.Tensor] = [ (i.contiguous() if isinstance(i, torch.Tensor) else torch.tensor(i).cuda()) @@ -981,7 +999,7 @@

    Source code for torch_tensorrt.dynamo.runtime._PythonTorchTensorRTModule

    ) if self.profiling_enabled: - import tempfile + import tempfile with tempfile.TemporaryDirectory() as tmpdir: self.cudagraph.debug_dump( @@ -1007,7 +1025,7 @@

    Source code for torch_tensorrt.dynamo.runtime._PythonTorchTensorRTModule

    return outputs
    -
    [docs] def enable_profiling(self, profiler: "trt.IProfiler" = None) -> None: +
    [docs] def enable_profiling(self, profiler: "trt.IProfiler" = None) -> None: """ Enable TensorRT profiling. After calling this function, TensorRT will report time spent on each layer in stdout for each forward run. @@ -1019,7 +1037,7 @@

    Source code for torch_tensorrt.dynamo.runtime._PythonTorchTensorRTModule

    self.profiling_enabled = True
    -
    [docs] def disable_profiling(self) -> None: +
    [docs] def disable_profiling(self) -> None: """ Disable TensorRT profiling. """ @@ -1029,7 +1047,7 @@

    Source code for torch_tensorrt.dynamo.runtime._PythonTorchTensorRTModule

    self.context = self.engine.create_execution_context() self.profiling_enabled = False
    -
    [docs] def get_layer_info(self) -> str: +
    [docs] def get_layer_info(self) -> str: """ Get layer info of the engine. Only support for TRT > 8.2. """ @@ -1039,7 +1057,7 @@

    Source code for torch_tensorrt.dynamo.runtime._PythonTorchTensorRTModule

    ) return engine_json
    -
    [docs] def validate_input_shapes(self, inputs: Sequence[torch.Tensor]) -> bool: +
    [docs] def validate_input_shapes(self, inputs: Sequence[torch.Tensor]) -> bool: """ Validates the input shapes of the forward function has changed """ @@ -1282,7 +1300,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1298,6 +1316,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1336,6 +1357,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1349,6 +1373,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_modules/torch_tensorrt/dynamo/runtime/_TorchTensorRTModule.html b/docs/_modules/torch_tensorrt/dynamo/runtime/_TorchTensorRTModule.html index 3a03a7d769..f1bd3ad0b5 100644 --- a/docs/_modules/torch_tensorrt/dynamo/runtime/_TorchTensorRTModule.html +++ b/docs/_modules/torch_tensorrt/dynamo/runtime/_TorchTensorRTModule.html @@ -9,7 +9,7 @@ - torch_tensorrt.dynamo.runtime._TorchTensorRTModule — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt.dynamo.runtime._TorchTensorRTModule — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -132,8 +132,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -154,6 +154,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -201,6 +204,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -216,7 +223,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -272,7 +290,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -467,23 +485,23 @@

    Source code for torch_tensorrt.dynamo.runtime._TorchTensorRTModule

    -from __future__ import annotations
    -
    -import base64
    -import copy
    -import logging
    -import pickle
    -from typing import Any, List, Optional, Tuple, Union
    -
    -import torch
    -from torch_tensorrt._Device import Device
    -from torch_tensorrt._enums import Platform
    -from torch_tensorrt._features import (
    +from __future__ import annotations
    +
    +import base64
    +import copy
    +import logging
    +import pickle
    +from typing import Any, List, Optional, Tuple, Union
    +
    +import torch
    +from torch_tensorrt._Device import Device
    +from torch_tensorrt._enums import Platform
    +from torch_tensorrt._features import (
         ENABLED_FEATURES,
         for_all_methods,
         needs_torch_tensorrt_runtime,
     )
    -from torch_tensorrt.dynamo._settings import CompilationSettings
    +from torch_tensorrt.dynamo._settings import CompilationSettings
     
     logger = logging.getLogger(__name__)
     
    @@ -519,7 +537,7 @@ 

    Source code for torch_tensorrt.dynamo.runtime._TorchTensorRTModule

    [docs]@for_all_methods(needs_torch_tensorrt_runtime) -class TorchTensorRTModule(torch.nn.Module): # type: ignore[misc] +class TorchTensorRTModule(torch.nn.Module): # type: ignore[misc] """TorchTensorRTModule is a PyTorch module which encompasses an arbitrary TensorRT Engine. This module is backed by the Torch-TensorRT runtime and is fully compatible with both @@ -539,7 +557,7 @@

    Source code for torch_tensorrt.dynamo.runtime._TorchTensorRTModule

    output_binding_names (List[str]): List of output TensorRT engine binding names in the order they should be returned """ - def __init__( + def __init__( self, serialized_engine: Optional[bytes] = None, input_binding_names: Optional[List[str]] = None, @@ -609,7 +627,7 @@

    Source code for torch_tensorrt.dynamo.runtime._TorchTensorRTModule

    ): self.setup_engine() - def _pack_engine_info(self) -> List[str | bytes]: + def _pack_engine_info(self) -> List[str | bytes]: target_device = ( self.settings.device if self.settings.device is not None @@ -643,16 +661,16 @@

    Source code for torch_tensorrt.dynamo.runtime._TorchTensorRTModule

    return engine_info - def get_streamable_device_memory_budget(self) -> Any: + def get_streamable_device_memory_budget(self) -> Any: return self.engine.streamable_device_memory_budget - def get_automatic_device_memory_budget(self) -> Any: + def get_automatic_device_memory_budget(self) -> Any: return self.engine.automatic_device_memory_budget - def get_device_memory_budget(self) -> Any: + def get_device_memory_budget(self) -> Any: return self.engine.device_memory_budget - def set_device_memory_budget(self, budget_bytes: int) -> int: + def set_device_memory_budget(self, budget_bytes: int) -> int: # Disable weight streaming for invalid budget size if budget_bytes < 0: budget_bytes = self.get_streamable_device_memory_budget() @@ -665,7 +683,7 @@

    Source code for torch_tensorrt.dynamo.runtime._TorchTensorRTModule

    return budget_bytes - def setup_engine(self) -> None: + def setup_engine(self) -> None: """ Setup engine for a module which has deferred engine setup. @@ -678,19 +696,19 @@

    Source code for torch_tensorrt.dynamo.runtime._TorchTensorRTModule

    return self.engine = torch.classes.tensorrt.Engine(self._pack_engine_info()) - def encode_metadata(self, metadata: Any) -> str: + def encode_metadata(self, metadata: Any) -> str: metadata = copy.deepcopy(metadata) dumped_metadata = pickle.dumps(metadata) encoded_metadata = base64.b64encode(dumped_metadata).decode("utf-8") return encoded_metadata @staticmethod - def decode_metadata(encoded_metadata: bytes) -> Any: + def decode_metadata(encoded_metadata: bytes) -> Any: dumped_metadata = base64.b64decode(encoded_metadata.encode("utf-8")) metadata = pickle.loads(dumped_metadata) return metadata - def get_extra_state(self) -> SerializedTorchTensorRTModuleFmt: + def get_extra_state(self) -> SerializedTorchTensorRTModuleFmt: if self.engine: return ( self.name, @@ -716,7 +734,7 @@

    Source code for torch_tensorrt.dynamo.runtime._TorchTensorRTModule

    self.output_binding_names, ) - def set_extra_state(self, state: SerializedTorchTensorRTModuleFmt) -> None: + def set_extra_state(self, state: SerializedTorchTensorRTModuleFmt) -> None: self.name = state[0] if state[1] is not None: @@ -741,10 +759,10 @@

    Source code for torch_tensorrt.dynamo.runtime._TorchTensorRTModule

    self.input_binding_names = state[2] self.output_binding_names = state[3] - def set_pre_allocated_outputs(self, enable: bool) -> None: + def set_pre_allocated_outputs(self, enable: bool) -> None: self.engine.use_pre_allocated_outputs = enable - def forward(self, *inputs: Any) -> torch.Tensor | Tuple[torch.Tensor, ...]: + def forward(self, *inputs: Any) -> torch.Tensor | Tuple[torch.Tensor, ...]: """Implementation of the forward pass for a TensorRT engine Args: @@ -779,7 +797,7 @@

    Source code for torch_tensorrt.dynamo.runtime._TorchTensorRTModule

    return tuple(outputs) - def enable_profiling(self, profiling_results_dir: Optional[str] = None) -> None: + def enable_profiling(self, profiling_results_dir: Optional[str] = None) -> None: """Enable the profiler to collect latency information about the execution of the engine Traces can be visualized using https://ui.perfetto.dev/ or compatible alternatives @@ -794,14 +812,14 @@

    Source code for torch_tensorrt.dynamo.runtime._TorchTensorRTModule

    self.engine.profile_path_prefix = profiling_results_dir self.engine.enable_profiling() - def disable_profiling(self) -> None: + def disable_profiling(self) -> None: """Disable the profiler""" if self.engine is None: raise RuntimeError("Engine has not been initialized yet.") self.engine.disable_profiling() - def get_layer_info(self) -> str: + def get_layer_info(self) -> str: """Get a JSON string containing the layer information encoded by the TensorRT engine in this module Returns: @@ -814,7 +832,7 @@

    Source code for torch_tensorrt.dynamo.runtime._TorchTensorRTModule

    layer_info: str = self.engine.get_engine_layer_info() return layer_info - def dump_layer_info(self) -> None: + def dump_layer_info(self) -> None: """Dump layer information encoded by the TensorRT engine in this module to STDOUT""" if self.engine is None: raise RuntimeError("Engine has not been initialized yet.") @@ -822,7 +840,7 @@

    Source code for torch_tensorrt.dynamo.runtime._TorchTensorRTModule

    self.engine.dump_engine_layer_info() @staticmethod - def _pack_binding_names(binding_names: List[str]) -> str: + def _pack_binding_names(binding_names: List[str]) -> str: delim = torch.ops.tensorrt.SERIALIZED_ENGINE_BINDING_DELIM()[0] packed_bindings: str = delim.join(binding_names) return packed_bindings
    @@ -1052,7 +1070,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1068,6 +1086,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1106,6 +1127,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1119,6 +1143,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_modules/torch_tensorrt/fx/fx2trt.html b/docs/_modules/torch_tensorrt/fx/fx2trt.html index feb0619d4c..a120552968 100644 --- a/docs/_modules/torch_tensorrt/fx/fx2trt.html +++ b/docs/_modules/torch_tensorrt/fx/fx2trt.html @@ -9,7 +9,7 @@ - torch_tensorrt.fx.fx2trt — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt.fx.fx2trt — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -132,8 +132,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -154,6 +154,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -201,6 +204,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -216,7 +223,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -272,7 +290,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -467,26 +485,26 @@

    Source code for torch_tensorrt.fx.fx2trt

    -import logging
    -import os
    -import warnings
    -from datetime import datetime
    -from typing import Any, Callable, Dict, List, NamedTuple, Optional, Sequence
    +import logging
    +import os
    +import warnings
    +from datetime import datetime
    +from typing import Any, Callable, Dict, List, NamedTuple, Optional, Sequence
     
    -import numpy
    +import numpy
     
     # @manual=//deeplearning/trt/python:py_tensorrt
    -import tensorrt as trt
    -import torch
    -import torch.fx
    -from torch._ops import OpOverload
    -from torch.fx.node import _get_qualified_name
    -from torch.fx.passes.shape_prop import TensorMetadata
    -
    -from .converter_registry import CONVERTERS
    -from .input_tensor_spec import InputTensorSpec
    -from .observer import Observer
    -from .utils import Frameworks, LowerPrecision, get_dynamic_dims, unified_dtype_converter
    +import tensorrt as trt
    +import torch
    +import torch.fx
    +from torch._ops import OpOverload
    +from torch.fx.node import _get_qualified_name
    +from torch.fx.passes.shape_prop import TensorMetadata
    +
    +from .converter_registry import CONVERTERS
    +from .input_tensor_spec import InputTensorSpec
    +from .observer import Observer
    +from .utils import Frameworks, LowerPrecision, get_dynamic_dims, unified_dtype_converter
     
     _LOGGER: logging.Logger = logging.getLogger(__name__)
     
    @@ -495,15 +513,15 @@ 

    Source code for torch_tensorrt.fx.fx2trt

     )
     
     
    -
    [docs]class TRTInterpreterResult(NamedTuple): +
    [docs]class TRTInterpreterResult(NamedTuple): engine: Any input_names: Sequence[str] output_names: Sequence[str] serialized_cache: bytearray
    -
    [docs]class TRTInterpreter(torch.fx.Interpreter): - def __init__( +
    [docs]class TRTInterpreter(torch.fx.Interpreter): + def __init__( self, module: torch.fx.GraphModule, input_specs: List[InputTensorSpec], @@ -548,7 +566,7 @@

    Source code for torch_tensorrt.fx.fx2trt

                 dict()
             )
     
    -    def validate_input_specs(self):
    +    def validate_input_specs(self):
             for shape, _, _, shape_ranges, has_batch_dim in self.input_specs:
                 if not self.network.has_implicit_batch_dimension:
                     assert (
    @@ -605,7 +623,7 @@ 

    Source code for torch_tensorrt.fx.fx2trt

                         len(shape_ranges) == 0
                     ), "shape_ranges are provided for input that doesn't have dynamic dim."
     
    -    def validate_conversion(self):
    +    def validate_conversion(self):
             missing_converter = set()
     
             for node in self.module.graph.nodes:
    @@ -621,7 +639,7 @@ 

    Source code for torch_tensorrt.fx.fx2trt

     
             return missing_converter
     
    -    def run(
    +    def run(
             self,
             max_batch_size=64,
             max_workspace_size=1 << 25,
    @@ -739,7 +757,7 @@ 

    Source code for torch_tensorrt.fx.fx2trt

                 engine, self._input_names, self._output_names, serialized_cache
             )
     
    -    def run_node(self, n):
    +    def run_node(self, n):
             self._cur_node_name = str(n)
             # add "_itensor_to_tensor_meta"
             kwargs = dict(n.kwargs)
    @@ -759,7 +777,7 @@ 

    Source code for torch_tensorrt.fx.fx2trt

     
             return trt_node
     
    -    def placeholder(self, target, args, kwargs):
    +    def placeholder(self, target, args, kwargs):
             self._input_names.append(target)
             shape, dtype, _, shape_ranges, has_batch_dim = self.input_specs[
                 self.input_specs_iter
    @@ -780,7 +798,7 @@ 

    Source code for torch_tensorrt.fx.fx2trt

                 dtype=unified_dtype_converter(dtype, Frameworks.TRT),
             )
     
    -    def call_module(self, target, args, kwargs):
    +    def call_module(self, target, args, kwargs):
             assert isinstance(target, str)
             submod = self.fetch_attr(target)
             submod_type = getattr(submod, "_base_class_origin", type(submod))
    @@ -794,7 +812,7 @@ 

    Source code for torch_tensorrt.fx.fx2trt

             assert self._cur_node_name is not None
             return converter(self.network, submod, args, kwargs, self._cur_node_name)
     
    -    def call_function(self, target, args, kwargs):
    +    def call_function(self, target, args, kwargs):
             converter = CONVERTERS.get(target)
             if not converter:
                 raise RuntimeError(
    @@ -804,7 +822,7 @@ 

    Source code for torch_tensorrt.fx.fx2trt

             assert self._cur_node_name is not None
             return converter(self.network, target, args, kwargs, self._cur_node_name)
     
    -    def call_method(self, target, args, kwargs):
    +    def call_method(self, target, args, kwargs):
             assert isinstance(target, str)
             converter = CONVERTERS.get(target)
     
    @@ -816,7 +834,7 @@ 

    Source code for torch_tensorrt.fx.fx2trt

             assert self._cur_node_name is not None
             return converter(self.network, target, args, kwargs, self._cur_node_name)
     
    -    def output(self, target, args, kwargs):
    +    def output(self, target, args, kwargs):
             assert len(args) == 1
             if isinstance(args[0], tuple):
                 outputs = args[0]
    @@ -1081,7 +1099,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1097,6 +1115,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1135,6 +1156,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1148,6 +1172,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_modules/torch_tensorrt/fx/input_tensor_spec.html b/docs/_modules/torch_tensorrt/fx/input_tensor_spec.html index 47c2744a6c..6ee9f30d79 100644 --- a/docs/_modules/torch_tensorrt/fx/input_tensor_spec.html +++ b/docs/_modules/torch_tensorrt/fx/input_tensor_spec.html @@ -9,7 +9,7 @@ - torch_tensorrt.fx.input_tensor_spec — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt.fx.input_tensor_spec — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -132,8 +132,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -154,6 +154,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -201,6 +204,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -216,7 +223,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -272,7 +290,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -467,15 +485,15 @@

    Source code for torch_tensorrt.fx.input_tensor_spec

    -from typing import Any, Iterable, List, NamedTuple, Optional, Sequence, Tuple
    +from typing import Any, Iterable, List, NamedTuple, Optional, Sequence, Tuple
     
    -import torch
    +import torch
     
    -from .types import Shape, ShapeRange
    -from .utils import get_dynamic_dims
    +from .types import Shape, ShapeRange
    +from .utils import get_dynamic_dims
     
     
    -def generate_input_specs(inputs, lower_setting, additional_inputs=None):
    +def generate_input_specs(inputs, lower_setting, additional_inputs=None):
         # dynamic_batch is TRT only flag.
         if (
             not lower_setting.explicit_batch_dimension
    @@ -534,7 +552,7 @@ 

    Source code for torch_tensorrt.fx.input_tensor_spec

    ) -
    [docs]class InputTensorSpec(NamedTuple): +
    [docs]class InputTensorSpec(NamedTuple): """ This class contains the information of a input tensor. @@ -564,7 +582,7 @@

    Source code for torch_tensorrt.fx.input_tensor_spec

    has_batch_dim: bool = True @classmethod - def from_tensor(cls, tensor: torch.Tensor) -> "InputTensorSpec": + def from_tensor(cls, tensor: torch.Tensor) -> "InputTensorSpec": """ Produce an InputTenosrSpec named tuple which contains the information of the given PyTorch tensor. @@ -578,7 +596,7 @@

    Source code for torch_tensorrt.fx.input_tensor_spec

    return cls(tensor.shape, tensor.dtype, tensor.device) @classmethod - def from_tensors(cls, tensors: Sequence[torch.Tensor]) -> List["InputTensorSpec"]: + def from_tensors(cls, tensors: Sequence[torch.Tensor]) -> List["InputTensorSpec"]: """ Produce a list of InputTenosrSpec named tuples which contain the information of all the given PyTorch tensors. @@ -593,7 +611,7 @@

    Source code for torch_tensorrt.fx.input_tensor_spec

    return [cls.from_tensor(t) for t in tensors] @classmethod - def from_tensors_with_dynamic_batch_size( + def from_tensors_with_dynamic_batch_size( cls, tensors: Sequence[torch.Tensor], batch_size_range: Tuple[int, int, int], @@ -647,7 +665,7 @@

    Source code for torch_tensorrt.fx.input_tensor_spec

    @classmethod # pyre-ignore [2]: Parameter `sample_input` must have a type other than `Any` - def find_batch_size_dim(cls, inputs: Any) -> []: + def find_batch_size_dim(cls, inputs: Any) -> []: if isinstance(inputs, torch.Tensor) or len(inputs) <= 1: return [0] shapes = [i.shape for i in inputs] @@ -686,7 +704,7 @@

    Source code for torch_tensorrt.fx.input_tensor_spec

    return bs_dim - def to_random_tensor(self, id=1): + def to_random_tensor(self, id=1): shape = tuple(self.shape) if len(get_dynamic_dims(shape)): # id=0 -> min shape @@ -699,7 +717,7 @@

    Source code for torch_tensorrt.fx.input_tensor_spec

    return torch.randn(shape).to(dtype=self.dtype, device=self.device) @staticmethod - def create_inputs_from_specs(input_specs: Iterable["InputTensorSpec"]): + def create_inputs_from_specs(input_specs: Iterable["InputTensorSpec"]): inputs = [] for spec in input_specs: inputs.append(spec.to_random_tensor()) @@ -707,7 +725,7 @@

    Source code for torch_tensorrt.fx.input_tensor_spec

    return inputs @staticmethod - def create_inputs_from_max_specs(input_specs: Iterable["InputTensorSpec"]): + def create_inputs_from_max_specs(input_specs: Iterable["InputTensorSpec"]): inputs = [] for spec in input_specs: inputs.append(spec.to_random_tensor(2)) @@ -939,7 +957,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -955,6 +973,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -993,6 +1014,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1006,6 +1030,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_modules/torch_tensorrt/fx/lower.html b/docs/_modules/torch_tensorrt/fx/lower.html index f7c15f508d..d341e87942 100644 --- a/docs/_modules/torch_tensorrt/fx/lower.html +++ b/docs/_modules/torch_tensorrt/fx/lower.html @@ -9,7 +9,7 @@ - torch_tensorrt.fx.lower — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt.fx.lower — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -132,8 +132,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -154,6 +154,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -201,6 +204,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -216,7 +223,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -272,7 +290,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -467,34 +485,34 @@

    Source code for torch_tensorrt.fx.lower

    -import dataclasses as dc
    -import logging
    -from typing import Any, Callable, Optional, Sequence
    +import dataclasses as dc
    +import logging
    +from typing import Any, Callable, Optional, Sequence
     
     # @manual=//deeplearning/trt/python:py_tensorrt
    -import tensorrt as trt
    -import torch
    -import torch.fx as fx
    -import torch.nn as nn
    -import torch_tensorrt.fx.tracer.dispatch_tracer.aten_tracer as aten_tracer
    -from torch.fx.passes.splitter_base import SplitResult
    -
    -from .fx2trt import TRTInterpreter, TRTInterpreterResult
    -from .lower_setting import LowerSetting
    -from .passes.lower_pass_manager_builder import LowerPassManagerBuilder
    -from .passes.pass_utils import PassFunc, validate_inference
    -from .tools.timing_cache_utils import TimingCacheManager
    -from .tools.trt_splitter import TRTSplitter, TRTSplitterSetting
    -from .tracer.acc_tracer import acc_tracer
    -from .trt_module import TRTModule
    -from .utils import LowerPrecision
    +import tensorrt as trt
    +import torch
    +import torch.fx as fx
    +import torch.nn as nn
    +import torch_tensorrt.fx.tracer.dispatch_tracer.aten_tracer as aten_tracer
    +from torch.fx.passes.splitter_base import SplitResult
    +
    +from .fx2trt import TRTInterpreter, TRTInterpreterResult
    +from .lower_setting import LowerSetting
    +from .passes.lower_pass_manager_builder import LowerPassManagerBuilder
    +from .passes.pass_utils import PassFunc, validate_inference
    +from .tools.timing_cache_utils import TimingCacheManager
    +from .tools.trt_splitter import TRTSplitter, TRTSplitterSetting
    +from .tracer.acc_tracer import acc_tracer
    +from .trt_module import TRTModule
    +from .utils import LowerPrecision
     
     logger = logging.getLogger(__name__)
     
     Input = Sequence[Any]
     
     
    -
    [docs]def compile( +
    [docs]def compile( module: nn.Module, input, min_acc_module_size: int = 10, @@ -559,18 +577,18 @@

    Source code for torch_tensorrt.fx.lower

     
     
     @dc.dataclass
    -class LowerTrtInterpreter:
    +class LowerTrtInterpreter:
         lower_setting: LowerSetting
         timing_cache_manager: TimingCacheManager
     
         @classmethod
    -    def create(cls, lower_setting):
    +    def create(cls, lower_setting):
             timing_cache_manager = TimingCacheManager(
                 lower_setting.timing_cache_prefix, lower_setting.save_timing_cache
             )
             return LowerTrtInterpreter(lower_setting, timing_cache_manager)
     
    -    def __call__(self, mod, input, split_name) -> TRTInterpreterResult:
    +    def __call__(self, mod, input, split_name) -> TRTInterpreterResult:
             assert self.lower_setting.input_specs, "Can't find input specs for lowering!"
             logger.info(
                 f"split_name={split_name}, input_specs={self.lower_setting.input_specs}"
    @@ -624,7 +642,7 @@ 

    Source code for torch_tensorrt.fx.lower

             return interp_result
     
     
    -def default_split_function(
    +def default_split_function(
         model: fx.GraphModule, inputs: Input, lower_setting: LowerSetting
     ) -> SplitResult:
         splitter_setting = TRTSplitterSetting()
    @@ -636,14 +654,14 @@ 

    Source code for torch_tensorrt.fx.lower

         return splitter.generate_split_results()
     
     
    -def create_lower_trt_interpreter(lower_setting: LowerSetting) -> LowerTrtInterpreter:
    +def create_lower_trt_interpreter(lower_setting: LowerSetting) -> LowerTrtInterpreter:
         return LowerTrtInterpreter.create(lower_setting)
     
     
    -def default_lower_pass(
    +def default_lower_pass(
         create_trt_interpreter: Callable[[LowerSetting], LowerTrtInterpreter],
     ) -> PassFunc:
    -    def lower_pass(
    +    def lower_pass(
             mod: nn.Module, input: Input, lower_setting: LowerSetting, module_name: str
         ) -> nn.Module:
             """
    @@ -653,10 +671,10 @@ 

    Source code for torch_tensorrt.fx.lower

             interpreter = create_trt_interpreter(lower_setting)
             interp_res: TRTInterpreterResult = interpreter(mod, input, module_name)
             if lower_setting.use_experimental_rt:
    -            import io
    +            import io
     
    -            from torch_tensorrt._Device import Device
    -            from torch_tensorrt.dynamo._TorchTensorRTModule import TorchTensorRTModule
    +            from torch_tensorrt._Device import Device
    +            from torch_tensorrt.dynamo._TorchTensorRTModule import TorchTensorRTModule
     
                 with io.BytesIO() as engine_bytes:
                     engine_bytes.write(interp_res.engine.serialize())
    @@ -685,7 +703,7 @@ 

    Source code for torch_tensorrt.fx.lower

     
     
     @dc.dataclass(frozen=True)
    -class Lowerer:
    +class Lowerer:
         """Lowers a module using fx2trt.
     
         This is a composable class to facilitate fx2trt. A normal fx2trt process
    @@ -709,7 +727,7 @@ 

    Source code for torch_tensorrt.fx.lower

         lower_pass_manager_builder: LowerPassManagerBuilder
     
         @classmethod
    -    def create(
    +    def create(
             cls,
             lower_setting: LowerSetting,
             interpreter_builder: Callable = create_lower_trt_interpreter,
    @@ -743,7 +761,7 @@ 

    Source code for torch_tensorrt.fx.lower

                     )
                 )
     
    -    def __call__(
    +    def __call__(
             self,
             module: nn.Module,
             inputs: Input,
    @@ -758,7 +776,7 @@ 

    Source code for torch_tensorrt.fx.lower

                 atol=atol,
                 rtol=rtol,
             )
    -        def do_lower(module: nn.Module, inputs: Input) -> nn.Module:
    +        def do_lower(module: nn.Module, inputs: Input) -> nn.Module:
                 module.eval()
                 if (
                     self.lower_pass_manager_builder.lower_setting.lower_precision
    @@ -1014,7 +1032,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1030,6 +1048,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1068,6 +1089,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1081,6 +1105,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_modules/torch_tensorrt/fx/trt_module.html b/docs/_modules/torch_tensorrt/fx/trt_module.html index 59a7ef0978..1bb6ded02d 100644 --- a/docs/_modules/torch_tensorrt/fx/trt_module.html +++ b/docs/_modules/torch_tensorrt/fx/trt_module.html @@ -9,7 +9,7 @@ - torch_tensorrt.fx.trt_module — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt.fx.trt_module — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -132,8 +132,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -154,6 +154,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -201,6 +204,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -216,7 +223,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -272,7 +290,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -467,17 +485,17 @@

    Source code for torch_tensorrt.fx.trt_module

    -from typing import Any, List, Sequence
    +from typing import Any, List, Sequence
     
     # @manual=//deeplearning/trt/python:py_tensorrt
    -import tensorrt as trt
    -import torch
    +import tensorrt as trt
    +import torch
     
    -from .utils import Frameworks, unified_dtype_converter
    +from .utils import Frameworks, unified_dtype_converter
     
     
    -
    [docs]class TRTModule(torch.nn.Module): - def __init__( +
    [docs]class TRTModule(torch.nn.Module): + def __init__( self, engine=None, input_names=None, output_names=None, cuda_graph_batch_size=-1 ): super(TRTModule, self).__init__() @@ -491,7 +509,7 @@

    Source code for torch_tensorrt.fx.trt_module

    if engine:
                 self._initialize()
     
    -    def _initialize(self):
    +    def _initialize(self):
             self.initialized = True
             self.context = self.engine.create_execution_context()
     
    @@ -560,18 +578,18 @@ 

    Source code for torch_tensorrt.fx.trt_module

    for idx in self.hidden_output_binding_indices_in_order
             ]
     
    -    def _check_initialized(self):
    +    def _check_initialized(self):
             if not self.initialized:
                 raise RuntimeError("TRTModule is not initialized.")
     
    -    def _on_state_dict(self, state_dict, prefix, local_metadata):
    +    def _on_state_dict(self, state_dict, prefix, local_metadata):
             self._check_initialized()
             state_dict[prefix + "engine"] = bytearray(self.engine.serialize())
             state_dict[prefix + "input_names"] = self.input_names
             state_dict[prefix + "output_names"] = self.output_names
             state_dict[prefix + "cuda_graph_batch_size"] = self.cuda_graph_batch_size
     
    -    def _load_from_state_dict(
    +    def _load_from_state_dict(
             self,
             state_dict,
             prefix,
    @@ -591,13 +609,13 @@ 

    Source code for torch_tensorrt.fx.trt_module

    self.output_names = state_dict[prefix + "output_names"]
             self._initialize()
     
    -    def __getstate__(self):
    +    def __getstate__(self):
             state = self.__dict__.copy()
             state["engine"] = bytearray(self.engine.serialize())
             state.pop("context", None)
             return state
     
    -    def __setstate__(self, state):
    +    def __setstate__(self, state):
             logger = trt.Logger()
             runtime = trt.Runtime(logger)
             state["engine"] = runtime.deserialize_cuda_engine(state["engine"])
    @@ -605,7 +623,7 @@ 

    Source code for torch_tensorrt.fx.trt_module

    if self.engine:
                 self.context = self.engine.create_execution_context()
     
    -    def forward(self, *inputs):
    +    def forward(self, *inputs):
             with torch.autograd.profiler.record_function("TRTModule:Forward"):
                 self._check_initialized()
     
    @@ -690,7 +708,7 @@ 

    Source code for torch_tensorrt.fx.trt_module

    return tuple(outputs)
     
    -    def enable_profiling(self, profiler: "trt.IProfiler" = None):
    +    def enable_profiling(self, profiler: "trt.IProfiler" = None):
             """
             Enable TensorRT profiling. After calling this function, TensorRT will report
             time spent on each layer in stdout for each forward run.
    @@ -700,7 +718,7 @@ 

    Source code for torch_tensorrt.fx.trt_module

    if not self.context.profiler:
                 self.context.profiler = trt.Profiler() if profiler is None else profiler
     
    -    def disable_profiling(self):
    +    def disable_profiling(self):
             """
             Disable TensorRT profiling.
             """
    @@ -710,7 +728,7 @@ 

    Source code for torch_tensorrt.fx.trt_module

    del self.context
             self.context = self.engine.create_execution_context()
     
    -    def get_layer_info(self) -> str:
    +    def get_layer_info(self) -> str:
             """
             Get layer info of the engine. Only support for TRT > 8.2.
             """
    @@ -942,7 +960,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -958,6 +976,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -996,6 +1017,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1009,6 +1033,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_modules/torch_tensorrt/logging.html b/docs/_modules/torch_tensorrt/logging.html index 5e9f31ad53..f2152f342c 100644 --- a/docs/_modules/torch_tensorrt/logging.html +++ b/docs/_modules/torch_tensorrt/logging.html @@ -9,7 +9,7 @@ - torch_tensorrt.logging — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt.logging — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -132,8 +132,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -154,6 +154,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -201,6 +204,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -216,7 +223,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -272,7 +290,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -467,24 +485,24 @@

    Source code for torch_tensorrt.logging

    -import logging
    -from typing import Any
    +import logging
    +from typing import Any
     
    -import torch
    -from torch_tensorrt._features import ENABLED_FEATURES
    +import torch
    +from torch_tensorrt._features import ENABLED_FEATURES
     
    -import tensorrt as trt
    +import tensorrt as trt
     
     logging.captureWarnings(True)
     _LOGGER = logging.getLogger("torch_tensorrt [TensorRT Conversion Context]")
     
     
    -class _TRTLogger(trt.ILogger):  # type: ignore[misc]
    +class _TRTLogger(trt.ILogger):  # type: ignore[misc]
     
    -    def __init__(self) -> None:
    +    def __init__(self) -> None:
             trt.ILogger.__init__(self)
     
    -    def log(self, severity: trt.ILogger.Severity, msg: str) -> None:
    +    def log(self, severity: trt.ILogger.Severity, msg: str) -> None:
             # TODO: Move to match once py39 reaches EoL
             if severity == trt.ILogger.Severity.INTERNAL_ERROR:
                 _LOGGER.critical(msg)
    @@ -502,7 +520,7 @@ 

    Source code for torch_tensorrt.logging

     TRT_LOGGER = _TRTLogger()
     
     
    -
    [docs]class internal_errors: +
    [docs]class internal_errors: """Context-manager to limit displayed log messages to just internal errors Example: @@ -513,12 +531,12 @@

    Source code for torch_tensorrt.logging

                     outputs = model_torchtrt(inputs)
         """
     
    -    def __enter__(self) -> None:
    +    def __enter__(self) -> None:
             self.external_lvl = _LOGGER.getEffectiveLevel()
             _LOGGER.setLevel(logging.CRITICAL)
     
             if ENABLED_FEATURES.torchscript_frontend:
    -            from torch_tensorrt.ts import logging as ts_logging
    +            from torch_tensorrt.ts import logging as ts_logging
     
                 self.ts_level = ts_logging.get_reportable_log_level()
                 ts_logging.set_reportable_log_level(ts_logging.Level.InternalError)
    @@ -529,11 +547,11 @@ 

    Source code for torch_tensorrt.logging

                     int(trt.ILogger.Severity.INTERNAL_ERROR)
                 )
     
    -    def __exit__(self, exc_type: Any, exc_value: Any, exc_tb: Any) -> None:
    +    def __exit__(self, exc_type: Any, exc_value: Any, exc_tb: Any) -> None:
             _LOGGER.setLevel(self.external_lvl)
     
             if ENABLED_FEATURES.torchscript_frontend:
    -            from torch_tensorrt.ts import logging as ts_logging
    +            from torch_tensorrt.ts import logging as ts_logging
     
                 ts_logging.set_reportable_log_level(self.ts_level)
     
    @@ -541,7 +559,7 @@ 

    Source code for torch_tensorrt.logging

                 torch.ops.tensorrt.set_logging_level(self.rt_level)
    -
    [docs]class errors: +
    [docs]class errors: """Context-manager to limit displayed log messages to just errors and above Example: @@ -552,12 +570,12 @@

    Source code for torch_tensorrt.logging

                     outputs = model_torchtrt(inputs)
         """
     
    -    def __enter__(self) -> None:
    +    def __enter__(self) -> None:
             self.external_lvl = _LOGGER.getEffectiveLevel()
             _LOGGER.setLevel(logging.ERROR)
     
             if ENABLED_FEATURES.torchscript_frontend:
    -            from torch_tensorrt.ts import logging as ts_logging
    +            from torch_tensorrt.ts import logging as ts_logging
     
                 self.ts_level = ts_logging.get_reportable_log_level()
                 ts_logging.set_reportable_log_level(ts_logging.Level.Error)
    @@ -566,11 +584,11 @@ 

    Source code for torch_tensorrt.logging

                 self.rt_level = torch.ops.tensorrt.get_logging_level()
                 torch.ops.tensorrt.set_logging_level(int(trt.ILogger.Severity.ERROR))
     
    -    def __exit__(self, exc_type: Any, exc_value: Any, exc_tb: Any) -> None:
    +    def __exit__(self, exc_type: Any, exc_value: Any, exc_tb: Any) -> None:
             _LOGGER.setLevel(self.external_lvl)
     
             if ENABLED_FEATURES.torchscript_frontend:
    -            from torch_tensorrt.ts import logging as ts_logging
    +            from torch_tensorrt.ts import logging as ts_logging
     
                 ts_logging.set_reportable_log_level(self.ts_level)
     
    @@ -578,7 +596,7 @@ 

    Source code for torch_tensorrt.logging

                 torch.ops.tensorrt.set_logging_level(self.rt_level)
    -
    [docs]class warnings: +
    [docs]class warnings: """Context-manager to limit displayed log messages to just warnings and above Example: @@ -589,12 +607,12 @@

    Source code for torch_tensorrt.logging

                     model_trt = torch_tensorrt.compile(model, **spec)
         """
     
    -    def __enter__(self) -> None:
    +    def __enter__(self) -> None:
             self.external_lvl = _LOGGER.getEffectiveLevel()
             _LOGGER.setLevel(logging.WARNING)
     
             if ENABLED_FEATURES.torchscript_frontend:
    -            from torch_tensorrt.ts import logging as ts_logging
    +            from torch_tensorrt.ts import logging as ts_logging
     
                 self.ts_level = ts_logging.get_reportable_log_level()
                 ts_logging.set_reportable_log_level(ts_logging.Level.Warning)
    @@ -603,11 +621,11 @@ 

    Source code for torch_tensorrt.logging

                 self.rt_level = torch.ops.tensorrt.get_logging_level()
                 torch.ops.tensorrt.set_logging_level(int(trt.ILogger.Severity.WARNING))
     
    -    def __exit__(self, exc_type: Any, exc_value: Any, exc_tb: Any) -> None:
    +    def __exit__(self, exc_type: Any, exc_value: Any, exc_tb: Any) -> None:
             _LOGGER.setLevel(self.external_lvl)
     
             if ENABLED_FEATURES.torchscript_frontend:
    -            from torch_tensorrt.ts import logging as ts_logging
    +            from torch_tensorrt.ts import logging as ts_logging
     
                 ts_logging.set_reportable_log_level(self.ts_level)
     
    @@ -615,7 +633,7 @@ 

    Source code for torch_tensorrt.logging

                 torch.ops.tensorrt.set_logging_level(self.rt_level)
    -
    [docs]class info: +
    [docs]class info: """Context-manager to display all info and greater severity messages Example: @@ -626,12 +644,12 @@

    Source code for torch_tensorrt.logging

                     model_trt = torch_tensorrt.compile(model, **spec)
         """
     
    -    def __enter__(self) -> None:
    +    def __enter__(self) -> None:
             self.external_lvl = _LOGGER.getEffectiveLevel()
             _LOGGER.setLevel(logging.INFO)
     
             if ENABLED_FEATURES.torchscript_frontend:
    -            from torch_tensorrt.ts import logging as ts_logging
    +            from torch_tensorrt.ts import logging as ts_logging
     
                 self.ts_level = ts_logging.get_reportable_log_level()
                 ts_logging.set_reportable_log_level(ts_logging.Level.Info)
    @@ -640,11 +658,11 @@ 

    Source code for torch_tensorrt.logging

                 self.rt_level = torch.ops.tensorrt.get_logging_level()
                 torch.ops.tensorrt.set_logging_level(int(trt.ILogger.Severity.INFO))
     
    -    def __exit__(self, exc_type: Any, exc_value: Any, exc_tb: Any) -> None:
    +    def __exit__(self, exc_type: Any, exc_value: Any, exc_tb: Any) -> None:
             _LOGGER.setLevel(self.external_lvl)
     
             if ENABLED_FEATURES.torchscript_frontend:
    -            from torch_tensorrt.ts import logging as ts_logging
    +            from torch_tensorrt.ts import logging as ts_logging
     
                 ts_logging.set_reportable_log_level(self.ts_level)
     
    @@ -652,7 +670,7 @@ 

    Source code for torch_tensorrt.logging

                 torch.ops.tensorrt.set_logging_level(self.rt_level)
    -
    [docs]class debug: +
    [docs]class debug: """Context-manager to display full debug information through the logger Example: @@ -663,12 +681,12 @@

    Source code for torch_tensorrt.logging

                     model_trt = torch_tensorrt.compile(model, **spec)
         """
     
    -    def __enter__(self) -> None:
    +    def __enter__(self) -> None:
             self.external_lvl = _LOGGER.getEffectiveLevel()
             _LOGGER.setLevel(logging.DEBUG)
     
             if ENABLED_FEATURES.torchscript_frontend:
    -            from torch_tensorrt.ts import logging as ts_logging
    +            from torch_tensorrt.ts import logging as ts_logging
     
                 self.ts_level = ts_logging.get_reportable_log_level()
                 ts_logging.set_reportable_log_level(ts_logging.Level.Debug)
    @@ -677,11 +695,11 @@ 

    Source code for torch_tensorrt.logging

                 self.rt_level = torch.ops.tensorrt.get_logging_level()
                 torch.ops.tensorrt.set_logging_level(int(trt.ILogger.Severity.VERBOSE))
     
    -    def __exit__(self, exc_type: Any, exc_value: Any, exc_tb: Any) -> None:
    +    def __exit__(self, exc_type: Any, exc_value: Any, exc_tb: Any) -> None:
             _LOGGER.setLevel(self.external_lvl)
     
             if ENABLED_FEATURES.torchscript_frontend:
    -            from torch_tensorrt.ts import logging as ts_logging
    +            from torch_tensorrt.ts import logging as ts_logging
     
                 ts_logging.set_reportable_log_level(self.ts_level)
     
    @@ -689,7 +707,7 @@ 

    Source code for torch_tensorrt.logging

                 torch.ops.tensorrt.set_logging_level(self.rt_level)
    -
    [docs]class graphs: +
    [docs]class graphs: """Context-manager to display the results of intermediate lowering passes as well as full debug information through the logger @@ -701,12 +719,12 @@

    Source code for torch_tensorrt.logging

                     model_trt = torch_tensorrt.compile(model, **spec)
         """
     
    -    def __enter__(self) -> None:
    +    def __enter__(self) -> None:
             self.external_lvl = _LOGGER.getEffectiveLevel()
             _LOGGER.setLevel(logging.NOTSET)
     
             if ENABLED_FEATURES.torchscript_frontend:
    -            from torch_tensorrt.ts import logging as ts_logging
    +            from torch_tensorrt.ts import logging as ts_logging
     
                 self.ts_level = ts_logging.get_reportable_log_level()
                 ts_logging.set_reportable_log_level(ts_logging.Level.Graph)
    @@ -715,11 +733,11 @@ 

    Source code for torch_tensorrt.logging

                 self.rt_level = torch.ops.tensorrt.get_logging_level()
                 torch.ops.tensorrt.set_logging_level(int(trt.ILogger.Severity.VERBOSE) + 1)
     
    -    def __exit__(self, exc_type: Any, exc_value: Any, exc_tb: Any) -> None:
    +    def __exit__(self, exc_type: Any, exc_value: Any, exc_tb: Any) -> None:
             _LOGGER.setLevel(self.external_lvl)
     
             if ENABLED_FEATURES.torchscript_frontend:
    -            from torch_tensorrt.ts import logging as ts_logging
    +            from torch_tensorrt.ts import logging as ts_logging
     
                 ts_logging.set_reportable_log_level(self.ts_level)
     
    @@ -951,7 +969,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -967,6 +985,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1005,6 +1026,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1018,6 +1042,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_modules/torch_tensorrt/runtime/_multi_device_safe_mode.html b/docs/_modules/torch_tensorrt/runtime/_multi_device_safe_mode.html index 1ae7163719..7d98e70c3a 100644 --- a/docs/_modules/torch_tensorrt/runtime/_multi_device_safe_mode.html +++ b/docs/_modules/torch_tensorrt/runtime/_multi_device_safe_mode.html @@ -9,7 +9,7 @@ - torch_tensorrt.runtime._multi_device_safe_mode — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt.runtime._multi_device_safe_mode — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -132,8 +132,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -154,6 +154,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -201,6 +204,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -216,7 +223,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -272,7 +290,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -467,11 +485,11 @@

    Source code for torch_tensorrt.runtime._multi_device_safe_mode

    -import logging
    -from typing import Any
    +import logging
    +from typing import Any
     
    -import torch
    -import torch_tensorrt
    +import torch
    +import torch_tensorrt
     
     if torch_tensorrt.ENABLED_FEATURES.torch_tensorrt_runtime:
         _PY_RT_MULTI_DEVICE_SAFE_MODE = torch.ops.tensorrt.get_multi_device_safe_mode()
    @@ -482,19 +500,19 @@ 

    Source code for torch_tensorrt.runtime._multi_device_safe_mode

    logger = logging.getLogger(__name__) -class _MultiDeviceSafeModeContextManager(object): +class _MultiDeviceSafeModeContextManager(object): """Helper class used in conjunction with `set_multi_device_safe_mode` Used to enable `set_multi_device_safe_mode` as a dual-purpose context manager """ - def __init__(self, old_mode: bool) -> None: + def __init__(self, old_mode: bool) -> None: self.old_mode = old_mode - def __enter__(self) -> "_MultiDeviceSafeModeContextManager": + def __enter__(self) -> "_MultiDeviceSafeModeContextManager": return self - def __exit__(self, *args: Any) -> None: + def __exit__(self, *args: Any) -> None: # Set multi-device safe mode back to old mode in Python global _PY_RT_MULTI_DEVICE_SAFE_MODE _PY_RT_MULTI_DEVICE_SAFE_MODE = self.old_mode @@ -504,7 +522,7 @@

    Source code for torch_tensorrt.runtime._multi_device_safe_mode

    torch.ops.tensorrt.set_multi_device_safe_mode(self.old_mode) -
    [docs]def set_multi_device_safe_mode(mode: bool) -> _MultiDeviceSafeModeContextManager: +
    [docs]def set_multi_device_safe_mode(mode: bool) -> _MultiDeviceSafeModeContextManager: """Sets the runtime (Python-only and default) into multi-device safe mode In the case that multiple devices are available on the system, in order for the @@ -762,7 +780,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -778,6 +796,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -816,6 +837,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -829,6 +853,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_modules/torch_tensorrt/ts/_compile_spec.html b/docs/_modules/torch_tensorrt/ts/_compile_spec.html index 8945cee3d8..58b2a8e438 100644 --- a/docs/_modules/torch_tensorrt/ts/_compile_spec.html +++ b/docs/_modules/torch_tensorrt/ts/_compile_spec.html @@ -9,7 +9,7 @@ - torch_tensorrt.ts._compile_spec — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt.ts._compile_spec — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -132,8 +132,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -154,6 +154,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -201,6 +204,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -216,7 +223,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -272,7 +290,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -467,25 +485,25 @@

    Source code for torch_tensorrt.ts._compile_spec

    -from __future__ import annotations
    +from __future__ import annotations
     
    -from copy import deepcopy
    -from typing import Any, Dict, List, Optional, Set
    +from copy import deepcopy
    +from typing import Any, Dict, List, Optional, Set
     
    -import torch
    -import torch_tensorrt._C.ts as _ts_C
    -from torch_tensorrt import _C
    -from torch_tensorrt._Device import Device
    -from torch_tensorrt._enums import DeviceType, EngineCapability, dtype
    -from torch_tensorrt._Input import Input
    -from torch_tensorrt.ts._Device import TorchScriptDevice
    -from torch_tensorrt.ts._Input import TorchScriptInput
    -from torch_tensorrt.ts.logging import Level, log
    +import torch
    +import torch_tensorrt._C.ts as _ts_C
    +from torch_tensorrt import _C
    +from torch_tensorrt._Device import Device
    +from torch_tensorrt._enums import DeviceType, EngineCapability, dtype
    +from torch_tensorrt._Input import Input
    +from torch_tensorrt.ts._Device import TorchScriptDevice
    +from torch_tensorrt.ts._Input import TorchScriptInput
    +from torch_tensorrt.ts.logging import Level, log
     
    -import tensorrt as trt
    +import tensorrt as trt
     
     
    -def _internal_input_to_torch_class_input(i: _C.Input) -> torch.classes.tensorrt._Input:
    +def _internal_input_to_torch_class_input(i: _C.Input) -> torch.classes.tensorrt._Input:
         clone = torch.classes.tensorrt._Input()
         clone._set_min(i.min)
         clone._set_opt(i.opt)
    @@ -498,7 +516,7 @@ 

    Source code for torch_tensorrt.ts._compile_spec

    < return clone -def _supported_input_size_type(input_size: Any) -> bool: +def _supported_input_size_type(input_size: Any) -> bool: if isinstance(input_size, torch.Size): return True elif isinstance(input_size, tuple): @@ -512,11 +530,11 @@

    Source code for torch_tensorrt.ts._compile_spec

    < ) -def _parse_op_precision(precision: Any) -> _C.dtype: +def _parse_op_precision(precision: Any) -> _C.dtype: return dtype._from(precision).to(_C.dtype) -def _parse_enabled_precisions(precisions: Any) -> Set[_C.dtype]: +def _parse_enabled_precisions(precisions: Any) -> Set[_C.dtype]: parsed_precisions = set() if any(isinstance(precisions, type) for type in [list, tuple, set]): for p in precisions: @@ -526,11 +544,11 @@

    Source code for torch_tensorrt.ts._compile_spec

    < return parsed_precisions -def _parse_device_type(device: Any) -> _C.DeviceType: +def _parse_device_type(device: Any) -> _C.DeviceType: return DeviceType._from(device).to(_C.DeviceType) -def _parse_device(device_info: Any) -> _C.Device: +def _parse_device(device_info: Any) -> _C.Device: if isinstance(device_info, dict): info = _C.Device() if "device_type" not in device_info: @@ -563,7 +581,7 @@

    Source code for torch_tensorrt.ts._compile_spec

    < ) -def _parse_torch_fallback(fallback_info: Dict[str, Any]) -> _ts_C.TorchFallback: +def _parse_torch_fallback(fallback_info: Dict[str, Any]) -> _ts_C.TorchFallback: info = _ts_C.TorchFallback() if "enabled" not in fallback_info: raise KeyError("Enabled is required parameter") @@ -585,7 +603,7 @@

    Source code for torch_tensorrt.ts._compile_spec

    < return info -def _parse_input_signature(input_signature: Any, depth: int = 0) -> Any: +def _parse_input_signature(input_signature: Any, depth: int = 0) -> Any: if depth > 2: raise AssertionError( "Input nesting depth exceeds max supported depth, use 1 level: [A, B], or 2 level: [A, (B, C)]" @@ -650,7 +668,7 @@

    Source code for torch_tensorrt.ts._compile_spec

    < ) -def _parse_compile_spec(compile_spec_: Dict[str, Any]) -> _ts_C.CompileSpec: +def _parse_compile_spec(compile_spec_: Dict[str, Any]) -> _ts_C.CompileSpec: # TODO: Use deepcopy to support partial compilation of collections compile_spec = deepcopy(compile_spec_) info = _ts_C.CompileSpec() @@ -773,7 +791,7 @@

    Source code for torch_tensorrt.ts._compile_spec

    < return info -
    [docs]def TensorRTCompileSpec( +
    [docs]def TensorRTCompileSpec( inputs: Optional[List[torch.Tensor | Input]] = None, input_signature: Optional[Any] = None, device: Optional[torch.device | Device] = None, @@ -1132,7 +1150,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1148,6 +1166,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1186,6 +1207,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1199,6 +1223,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_modules/torch_tensorrt/ts/_compiler.html b/docs/_modules/torch_tensorrt/ts/_compiler.html index 34b655dbfe..189edbd4d1 100644 --- a/docs/_modules/torch_tensorrt/ts/_compiler.html +++ b/docs/_modules/torch_tensorrt/ts/_compiler.html @@ -9,7 +9,7 @@ - torch_tensorrt.ts._compiler — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt.ts._compiler — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -132,8 +132,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -154,6 +154,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -201,6 +204,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -216,7 +223,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -272,7 +290,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -467,19 +485,19 @@

    Source code for torch_tensorrt.ts._compiler

    -from __future__ import annotations
    +from __future__ import annotations
     
    -from typing import Any, List, Optional, Sequence, Set, Tuple
    +from typing import Any, List, Optional, Sequence, Set, Tuple
     
    -import torch
    -import torch_tensorrt._C.ts as _C
    -from torch_tensorrt._Device import Device
    -from torch_tensorrt._enums import EngineCapability, dtype
    -from torch_tensorrt._Input import Input
    -from torch_tensorrt.ts._compile_spec import _parse_compile_spec, _parse_device
    +import torch
    +import torch_tensorrt._C.ts as _C
    +from torch_tensorrt._Device import Device
    +from torch_tensorrt._enums import EngineCapability, dtype
    +from torch_tensorrt._Input import Input
    +from torch_tensorrt.ts._compile_spec import _parse_compile_spec, _parse_device
     
     
    -
    [docs]def compile( +
    [docs]def compile( module: torch.jit.ScriptModule, inputs: Optional[Sequence[Input | torch.Tensor]] = None, input_signature: Optional[Tuple[Input | torch.Tensor | Sequence[Any]]] = None, @@ -627,7 +645,7 @@

    Source code for torch_tensorrt.ts._compiler

         return compiled_module
    -
    [docs]def convert_method_to_trt_engine( +
    [docs]def convert_method_to_trt_engine( module: torch.jit.ScriptModule, method_name: str = "forward", inputs: Optional[Sequence[Input | torch.Tensor]] = None, @@ -739,7 +757,7 @@

    Source code for torch_tensorrt.ts._compiler

             module._c, method_name, _parse_compile_spec(compile_spec)
         )
     
    -    import io
    +    import io
     
         with io.BytesIO() as engine_bytes:
             engine_bytes.write(engine_str)
    @@ -748,7 +766,7 @@ 

    Source code for torch_tensorrt.ts._compiler

         return engine_bytearray
    -
    [docs]def embed_engine_in_new_module( +
    [docs]def embed_engine_in_new_module( serialized_engine: bytes, input_binding_names: Optional[List[str]] = None, output_binding_names: Optional[List[str]] = None, @@ -794,7 +812,7 @@

    Source code for torch_tensorrt.ts._compiler

         return wrapped_mod
    -
    [docs]def check_method_op_support( +
    [docs]def check_method_op_support( module: torch.jit.ScriptModule, method_name: str = "forward" ) -> bool: """Checks to see if a method is fully supported by torch_tensorrt @@ -1038,7 +1056,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1054,6 +1072,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1092,6 +1113,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1105,6 +1129,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_modules/torch_tensorrt/ts/ptq.html b/docs/_modules/torch_tensorrt/ts/ptq.html index d8bca52250..9ab2d1112d 100644 --- a/docs/_modules/torch_tensorrt/ts/ptq.html +++ b/docs/_modules/torch_tensorrt/ts/ptq.html @@ -9,7 +9,7 @@ - torch_tensorrt.ts.ptq — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt.ts.ptq — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -132,8 +132,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -154,6 +154,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -201,6 +204,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -216,7 +223,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -272,7 +290,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -467,38 +485,38 @@

    Source code for torch_tensorrt.ts.ptq

    -import sys
    -from typing import Any, List, Optional
    +import sys
    +from typing import Any, List, Optional
     
     if sys.version_info >= (3, 11):
    -    from typing import Self
    +    from typing import Self
     else:
    -    from typing_extensions import Self
    +    from typing_extensions import Self
     
    -import os
    -from enum import Enum
    +import os
    +from enum import Enum
     
    -import torch
    -from torch_tensorrt import _C
    -from torch_tensorrt.ts.logging import Level, log
    +import torch
    +from torch_tensorrt import _C
    +from torch_tensorrt.ts.logging import Level, log
     
     
    -
    [docs]class CalibrationAlgo(Enum): +
    [docs]class CalibrationAlgo(Enum): ENTROPY_CALIBRATION = _C.CalibrationAlgo.ENTROPY_CALIBRATION ENTROPY_CALIBRATION_2 = _C.CalibrationAlgo.ENTROPY_CALIBRATION_2 LEGACY_CALIBRATION = _C.CalibrationAlgo.LEGACY_CALIBRATION MINMAX_CALIBRATION = _C.CalibrationAlgo.MINMAX_CALIBRATION
    -def get_cache_mode_batch(self: object) -> None: +def get_cache_mode_batch(self: object) -> None: return None -def get_batch_size(self: object) -> int: +def get_batch_size(self: object) -> int: return 1 -def get_batch(self: object, _: Any) -> Optional[List[int]]: +def get_batch(self: object, _: Any) -> Optional[List[int]]: if self.current_batch_idx + self.batch_size > len(self.data_loader.dataset): return None @@ -513,7 +531,7 @@

    Source code for torch_tensorrt.ts.ptq

         return inputs_gpu
     
     
    -def read_calibration_cache(self: object) -> bytes:
    +def read_calibration_cache(self: object) -> bytes:
         if self.cache_file and self.use_cache:
             if os.path.exists(self.cache_file):
                 with open(self.cache_file, "rb") as f:
    @@ -525,7 +543,7 @@ 

    Source code for torch_tensorrt.ts.ptq

             return b""
     
     
    -def write_calibration_cache(self: object, cache: bytes) -> None:
    +def write_calibration_cache(self: object, cache: bytes) -> None:
         if self.cache_file:
             with open(self.cache_file, "wb") as f:
                 f.write(cache)
    @@ -536,11 +554,11 @@ 

    Source code for torch_tensorrt.ts.ptq

     # deepcopy (which involves pickling) is performed on the compile_spec internally during compilation.
     # We register this __reduce__ function for pickler to identity the calibrator object returned by DataLoaderCalibrator during deepcopy.
     # This should be the object's local name relative to the module https://docs.python.org/3/library/pickle.html#object.__reduce__
    -def __reduce__(self: object) -> str:
    +def __reduce__(self: object) -> str:
         return self.__class__.__name__
     
     
    -
    [docs]class DataLoaderCalibrator(object): +
    [docs]class DataLoaderCalibrator(object): """ Constructs a calibrator class in TensorRT and uses pytorch dataloader to load/preprocess data which is passed during calibration. @@ -553,10 +571,10 @@

    Source code for torch_tensorrt.ts.ptq

             device (Device): device on which calibration data is copied to.
         """
     
    -    def __init__(self, **kwargs: Any):
    +    def __init__(self, **kwargs: Any):
             pass
     
    -    def __new__(cls, *args: Any, **kwargs: Any) -> Self:
    +    def __new__(cls, *args: Any, **kwargs: Any) -> Self:
             dataloader = args[0]
             algo_type = kwargs.get("algo_type", CalibrationAlgo.ENTROPY_CALIBRATION_2)
             cache_file = kwargs.get("cache_file", None)
    @@ -631,7 +649,7 @@ 

    Source code for torch_tensorrt.ts.ptq

                 )
    -
    [docs]class CacheCalibrator(object): +
    [docs]class CacheCalibrator(object): """ Constructs a calibrator class in TensorRT which directly uses pre-existing cache file for calibration. @@ -640,10 +658,10 @@

    Source code for torch_tensorrt.ts.ptq

             algo_type (CalibrationAlgo): choice of calibration algorithm.
         """
     
    -    def __init__(self, **kwargs: Any):
    +    def __init__(self, **kwargs: Any):
             pass
     
    -    def __new__(cls, *args: Any, **kwargs: Any) -> Self:
    +    def __new__(cls, *args: Any, **kwargs: Any) -> Self:
             cache_file = args[0]
             algo_type = kwargs.get("algo_type", CalibrationAlgo.ENTROPY_CALIBRATION_2)
     
    @@ -915,7 +933,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -931,6 +949,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -969,6 +990,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -982,6 +1006,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/_sources/user_guide/saving_models.rst.txt b/docs/_sources/user_guide/saving_models.rst.txt index 42cff7b954..dc4b5da222 100644 --- a/docs/_sources/user_guide/saving_models.rst.txt +++ b/docs/_sources/user_guide/saving_models.rst.txt @@ -34,7 +34,7 @@ Here's an example usage model = MyModel().eval().cuda() inputs = [torch.randn((1, 3, 224, 224)).cuda()] # trt_ep is a torch.fx.GraphModule object - trt_gm = torch_tensorrt.compile(model, ir="dynamo", inputs) + trt_gm = torch_tensorrt.compile(model, ir="dynamo", inputs=inputs) torch_tensorrt.save(trt_gm, "trt.ep", inputs=inputs) # Later, you can load it and run inference @@ -52,7 +52,7 @@ b) Torchscript model = MyModel().eval().cuda() inputs = [torch.randn((1, 3, 224, 224)).cuda()] # trt_gm is a torch.fx.GraphModule object - trt_gm = torch_tensorrt.compile(model, ir="dynamo", inputs) + trt_gm = torch_tensorrt.compile(model, ir="dynamo", inputs=inputs) torch_tensorrt.save(trt_gm, "trt.ts", output_format="torchscript", inputs=inputs) # Later, you can load it and run inference @@ -73,7 +73,7 @@ For `ir=ts`, this behavior stays the same in 2.X versions as well. model = MyModel().eval().cuda() inputs = [torch.randn((1, 3, 224, 224)).cuda()] - trt_ts = torch_tensorrt.compile(model, ir="ts", inputs) # Output is a ScriptModule object + trt_ts = torch_tensorrt.compile(model, ir="ts", inputs=inputs) # Output is a ScriptModule object torch.jit.save(trt_ts, "trt_model.ts") # Later, you can load it and run inference diff --git a/docs/_static/documentation_options.js b/docs/_static/documentation_options.js index 5bf9b066f8..efbb769960 100644 --- a/docs/_static/documentation_options.js +++ b/docs/_static/documentation_options.js @@ -1,6 +1,6 @@ var DOCUMENTATION_OPTIONS = { URL_ROOT: document.getElementById("documentation_options").getAttribute('data-url_root'), - VERSION: 'v2.6.0.dev0+70e2a38', + VERSION: 'v2.7.0.dev0+d6be4ba', LANGUAGE: 'en', COLLAPSE_INDEX: false, BUILDER: 'html', diff --git a/docs/_static/js/vendor/anchor.min.js b/docs/_static/js/vendor/anchor.min.js index 1216eeac26..1c2b86faed 100644 --- a/docs/_static/js/vendor/anchor.min.js +++ b/docs/_static/js/vendor/anchor.min.js @@ -1,9 +1,9 @@ // @license magnet:?xt=urn:btih:d3d9a9a6595521f9666a5e94cc830dab83b65699&dn=expat.txt Expat // -// AnchorJS - v4.3.0 - 2020-10-21 +// AnchorJS - v4.3.1 - 2021-04-17 // https://www.bryanbraun.com/anchorjs/ -// Copyright (c) 2020 Bryan Braun; Licensed MIT +// Copyright (c) 2021 Bryan Braun; Licensed MIT // // @license magnet:?xt=urn:btih:d3d9a9a6595521f9666a5e94cc830dab83b65699&dn=expat.txt Expat -!function(A,e){"use strict";"function"==typeof define&&define.amd?define([],e):"object"==typeof module&&module.exports?module.exports=e():(A.AnchorJS=e(),A.anchors=new A.AnchorJS)}(this,function(){"use strict";return function(A){function d(A){A.icon=Object.prototype.hasOwnProperty.call(A,"icon")?A.icon:"",A.visible=Object.prototype.hasOwnProperty.call(A,"visible")?A.visible:"hover",A.placement=Object.prototype.hasOwnProperty.call(A,"placement")?A.placement:"right",A.ariaLabel=Object.prototype.hasOwnProperty.call(A,"ariaLabel")?A.ariaLabel:"Anchor",A.class=Object.prototype.hasOwnProperty.call(A,"class")?A.class:"",A.base=Object.prototype.hasOwnProperty.call(A,"base")?A.base:"",A.truncate=Object.prototype.hasOwnProperty.call(A,"truncate")?Math.floor(A.truncate):64,A.titleText=Object.prototype.hasOwnProperty.call(A,"titleText")?A.titleText:""}function f(A){var e;if("string"==typeof A||A instanceof String)e=[].slice.call(document.querySelectorAll(A));else{if(!(Array.isArray(A)||A instanceof NodeList))throw new TypeError("The selector provided to AnchorJS was invalid.");e=[].slice.call(A)}return e}this.options=A||{},this.elements=[],d(this.options),this.isTouchDevice=function(){return Boolean("ontouchstart"in window||window.TouchEvent||window.DocumentTouch&&document instanceof DocumentTouch)},this.add=function(A){var e,t,o,n,i,s,a,r,c,l,h,u,p=[];if(d(this.options),"touch"===(h=this.options.visible)&&(h=this.isTouchDevice()?"always":"hover"),0===(e=f(A=A||"h2, h3, h4, h5, h6")).length)return this;for(!function(){if(null!==document.head.querySelector("style.anchorjs"))return;var A,e=document.createElement("style");e.className="anchorjs",e.appendChild(document.createTextNode("")),void 0===(A=document.head.querySelector('[rel="stylesheet"],style'))?document.head.appendChild(e):document.head.insertBefore(e,A);e.sheet.insertRule(".anchorjs-link{opacity:0;text-decoration:none;-webkit-font-smoothing:antialiased;-moz-osx-font-smoothing:grayscale}",e.sheet.cssRules.length),e.sheet.insertRule(":hover>.anchorjs-link,.anchorjs-link:focus{opacity:1}",e.sheet.cssRules.length),e.sheet.insertRule("[data-anchorjs-icon]::after{content:attr(data-anchorjs-icon)}",e.sheet.cssRules.length),e.sheet.insertRule('@font-face{font-family:anchorjs-icons;src:url(data:n/a;base64,AAEAAAALAIAAAwAwT1MvMg8yG2cAAAE4AAAAYGNtYXDp3gC3AAABpAAAAExnYXNwAAAAEAAAA9wAAAAIZ2x5ZlQCcfwAAAH4AAABCGhlYWQHFvHyAAAAvAAAADZoaGVhBnACFwAAAPQAAAAkaG10eASAADEAAAGYAAAADGxvY2EACACEAAAB8AAAAAhtYXhwAAYAVwAAARgAAAAgbmFtZQGOH9cAAAMAAAAAunBvc3QAAwAAAAADvAAAACAAAQAAAAEAAHzE2p9fDzz1AAkEAAAAAADRecUWAAAAANQA6R8AAAAAAoACwAAAAAgAAgAAAAAAAAABAAADwP/AAAACgAAA/9MCrQABAAAAAAAAAAAAAAAAAAAAAwABAAAAAwBVAAIAAAAAAAIAAAAAAAAAAAAAAAAAAAAAAAMCQAGQAAUAAAKZAswAAACPApkCzAAAAesAMwEJAAAAAAAAAAAAAAAAAAAAARAAAAAAAAAAAAAAAAAAAAAAQAAg//0DwP/AAEADwABAAAAAAQAAAAAAAAAAAAAAIAAAAAAAAAIAAAACgAAxAAAAAwAAAAMAAAAcAAEAAwAAABwAAwABAAAAHAAEADAAAAAIAAgAAgAAACDpy//9//8AAAAg6cv//f///+EWNwADAAEAAAAAAAAAAAAAAAAACACEAAEAAAAAAAAAAAAAAAAxAAACAAQARAKAAsAAKwBUAAABIiYnJjQ3NzY2MzIWFxYUBwcGIicmNDc3NjQnJiYjIgYHBwYUFxYUBwYGIwciJicmNDc3NjIXFhQHBwYUFxYWMzI2Nzc2NCcmNDc2MhcWFAcHBgYjARQGDAUtLXoWOR8fORYtLTgKGwoKCjgaGg0gEhIgDXoaGgkJBQwHdR85Fi0tOAobCgoKOBoaDSASEiANehoaCQkKGwotLXoWOR8BMwUFLYEuehYXFxYugC44CQkKGwo4GkoaDQ0NDXoaShoKGwoFBe8XFi6ALjgJCQobCjgaShoNDQ0NehpKGgobCgoKLYEuehYXAAAADACWAAEAAAAAAAEACAAAAAEAAAAAAAIAAwAIAAEAAAAAAAMACAAAAAEAAAAAAAQACAAAAAEAAAAAAAUAAQALAAEAAAAAAAYACAAAAAMAAQQJAAEAEAAMAAMAAQQJAAIABgAcAAMAAQQJAAMAEAAMAAMAAQQJAAQAEAAMAAMAAQQJAAUAAgAiAAMAAQQJAAYAEAAMYW5jaG9yanM0MDBAAGEAbgBjAGgAbwByAGoAcwA0ADAAMABAAAAAAwAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAABAAH//wAP) format("truetype")}',e.sheet.cssRules.length)}(),t=document.querySelectorAll("[id]"),o=[].map.call(t,function(A){return A.id}),i=0;i\]./()*\\\n\t\b\v\u00A0]/g,"-").replace(/-{2,}/g,"-").substring(0,this.options.truncate).replace(/^-+|-+$/gm,"").toLowerCase()},this.hasAnchorJSLink=function(A){var e=A.firstChild&&-1<(" "+A.firstChild.className+" ").indexOf(" anchorjs-link "),t=A.lastChild&&-1<(" "+A.lastChild.className+" ").indexOf(" anchorjs-link ");return e||t||!1}}}); +!function(A,e){"use strict";"function"==typeof define&&define.amd?define([],e):"object"==typeof module&&module.exports?module.exports=e():(A.AnchorJS=e(),A.anchors=new A.AnchorJS)}(this,function(){"use strict";return function(A){function d(A){A.icon=Object.prototype.hasOwnProperty.call(A,"icon")?A.icon:"",A.visible=Object.prototype.hasOwnProperty.call(A,"visible")?A.visible:"hover",A.placement=Object.prototype.hasOwnProperty.call(A,"placement")?A.placement:"right",A.ariaLabel=Object.prototype.hasOwnProperty.call(A,"ariaLabel")?A.ariaLabel:"Anchor",A.class=Object.prototype.hasOwnProperty.call(A,"class")?A.class:"",A.base=Object.prototype.hasOwnProperty.call(A,"base")?A.base:"",A.truncate=Object.prototype.hasOwnProperty.call(A,"truncate")?Math.floor(A.truncate):64,A.titleText=Object.prototype.hasOwnProperty.call(A,"titleText")?A.titleText:""}function w(A){var e;if("string"==typeof A||A instanceof String)e=[].slice.call(document.querySelectorAll(A));else{if(!(Array.isArray(A)||A instanceof NodeList))throw new TypeError("The selector provided to AnchorJS was invalid.");e=[].slice.call(A)}return e}this.options=A||{},this.elements=[],d(this.options),this.isTouchDevice=function(){return Boolean("ontouchstart"in window||window.TouchEvent||window.DocumentTouch&&document instanceof DocumentTouch)},this.add=function(A){var e,t,o,i,n,s,a,c,r,l,h,u,p=[];if(d(this.options),"touch"===(l=this.options.visible)&&(l=this.isTouchDevice()?"always":"hover"),0===(e=w(A=A||"h2, h3, h4, h5, h6")).length)return this;for(null===document.head.querySelector("style.anchorjs")&&((u=document.createElement("style")).className="anchorjs",u.appendChild(document.createTextNode("")),void 0===(A=document.head.querySelector('[rel="stylesheet"],style'))?document.head.appendChild(u):document.head.insertBefore(u,A),u.sheet.insertRule(".anchorjs-link{opacity:0;text-decoration:none;-webkit-font-smoothing:antialiased;-moz-osx-font-smoothing:grayscale}",u.sheet.cssRules.length),u.sheet.insertRule(":hover>.anchorjs-link,.anchorjs-link:focus{opacity:1}",u.sheet.cssRules.length),u.sheet.insertRule("[data-anchorjs-icon]::after{content:attr(data-anchorjs-icon)}",u.sheet.cssRules.length),u.sheet.insertRule('@font-face{font-family:anchorjs-icons;src:url(data:n/a;base64,AAEAAAALAIAAAwAwT1MvMg8yG2cAAAE4AAAAYGNtYXDp3gC3AAABpAAAAExnYXNwAAAAEAAAA9wAAAAIZ2x5ZlQCcfwAAAH4AAABCGhlYWQHFvHyAAAAvAAAADZoaGVhBnACFwAAAPQAAAAkaG10eASAADEAAAGYAAAADGxvY2EACACEAAAB8AAAAAhtYXhwAAYAVwAAARgAAAAgbmFtZQGOH9cAAAMAAAAAunBvc3QAAwAAAAADvAAAACAAAQAAAAEAAHzE2p9fDzz1AAkEAAAAAADRecUWAAAAANQA6R8AAAAAAoACwAAAAAgAAgAAAAAAAAABAAADwP/AAAACgAAA/9MCrQABAAAAAAAAAAAAAAAAAAAAAwABAAAAAwBVAAIAAAAAAAIAAAAAAAAAAAAAAAAAAAAAAAMCQAGQAAUAAAKZAswAAACPApkCzAAAAesAMwEJAAAAAAAAAAAAAAAAAAAAARAAAAAAAAAAAAAAAAAAAAAAQAAg//0DwP/AAEADwABAAAAAAQAAAAAAAAAAAAAAIAAAAAAAAAIAAAACgAAxAAAAAwAAAAMAAAAcAAEAAwAAABwAAwABAAAAHAAEADAAAAAIAAgAAgAAACDpy//9//8AAAAg6cv//f///+EWNwADAAEAAAAAAAAAAAAAAAAACACEAAEAAAAAAAAAAAAAAAAxAAACAAQARAKAAsAAKwBUAAABIiYnJjQ3NzY2MzIWFxYUBwcGIicmNDc3NjQnJiYjIgYHBwYUFxYUBwYGIwciJicmNDc3NjIXFhQHBwYUFxYWMzI2Nzc2NCcmNDc2MhcWFAcHBgYjARQGDAUtLXoWOR8fORYtLTgKGwoKCjgaGg0gEhIgDXoaGgkJBQwHdR85Fi0tOAobCgoKOBoaDSASEiANehoaCQkKGwotLXoWOR8BMwUFLYEuehYXFxYugC44CQkKGwo4GkoaDQ0NDXoaShoKGwoFBe8XFi6ALjgJCQobCjgaShoNDQ0NehpKGgobCgoKLYEuehYXAAAADACWAAEAAAAAAAEACAAAAAEAAAAAAAIAAwAIAAEAAAAAAAMACAAAAAEAAAAAAAQACAAAAAEAAAAAAAUAAQALAAEAAAAAAAYACAAAAAMAAQQJAAEAEAAMAAMAAQQJAAIABgAcAAMAAQQJAAMAEAAMAAMAAQQJAAQAEAAMAAMAAQQJAAUAAgAiAAMAAQQJAAYAEAAMYW5jaG9yanM0MDBAAGEAbgBjAGgAbwByAGoAcwA0ADAAMABAAAAAAwAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAABAAH//wAP) format("truetype")}',u.sheet.cssRules.length)),u=document.querySelectorAll("[id]"),t=[].map.call(u,function(A){return A.id}),i=0;i\]./()*\\\n\t\b\v\u00A0]/g,"-").replace(/-{2,}/g,"-").substring(0,this.options.truncate).replace(/^-+|-+$/gm,"").toLowerCase()},this.hasAnchorJSLink=function(A){var e=A.firstChild&&-1<(" "+A.firstChild.className+" ").indexOf(" anchorjs-link "),A=A.lastChild&&-1<(" "+A.lastChild.className+" ").indexOf(" anchorjs-link ");return e||A||!1}}}); // @license-end \ No newline at end of file diff --git a/docs/_static/pygments.css b/docs/_static/pygments.css index 0d49244eda..5f2b0a250a 100644 --- a/docs/_static/pygments.css +++ b/docs/_static/pygments.css @@ -6,26 +6,26 @@ span.linenos.special { color: #000000; background-color: #ffffc0; padding-left: .highlight .hll { background-color: #ffffcc } .highlight { background: #eeffcc; } .highlight .c { color: #408090; font-style: italic } /* Comment */ -.highlight .err { border: 1px solid #FF0000 } /* Error */ +.highlight .err { border: 1px solid #F00 } /* Error */ .highlight .k { color: #007020; font-weight: bold } /* Keyword */ -.highlight .o { color: #666666 } /* Operator */ +.highlight .o { color: #666 } /* Operator */ .highlight .ch { color: #408090; font-style: italic } /* Comment.Hashbang */ .highlight .cm { color: #408090; font-style: italic } /* Comment.Multiline */ .highlight .cp { color: #007020 } /* Comment.Preproc */ .highlight .cpf { color: #408090; font-style: italic } /* Comment.PreprocFile */ .highlight .c1 { color: #408090; font-style: italic } /* Comment.Single */ -.highlight .cs { color: #408090; background-color: #fff0f0 } /* Comment.Special */ +.highlight .cs { color: #408090; background-color: #FFF0F0 } /* Comment.Special */ .highlight .gd { color: #A00000 } /* Generic.Deleted */ .highlight .ge { font-style: italic } /* Generic.Emph */ .highlight .ges { font-weight: bold; font-style: italic } /* Generic.EmphStrong */ -.highlight .gr { color: #FF0000 } /* Generic.Error */ +.highlight .gr { color: #F00 } /* Generic.Error */ .highlight .gh { color: #000080; font-weight: bold } /* Generic.Heading */ .highlight .gi { color: #00A000 } /* Generic.Inserted */ -.highlight .go { color: #333333 } /* Generic.Output */ -.highlight .gp { color: #c65d09; font-weight: bold } /* Generic.Prompt */ +.highlight .go { color: #333 } /* Generic.Output */ +.highlight .gp { color: #C65D09; font-weight: bold } /* Generic.Prompt */ .highlight .gs { font-weight: bold } /* Generic.Strong */ .highlight .gu { color: #800080; font-weight: bold } /* Generic.Subheading */ -.highlight .gt { color: #0044DD } /* Generic.Traceback */ +.highlight .gt { color: #04D } /* Generic.Traceback */ .highlight .kc { color: #007020; font-weight: bold } /* Keyword.Constant */ .highlight .kd { color: #007020; font-weight: bold } /* Keyword.Declaration */ .highlight .kn { color: #007020; font-weight: bold } /* Keyword.Namespace */ @@ -33,43 +33,43 @@ span.linenos.special { color: #000000; background-color: #ffffc0; padding-left: .highlight .kr { color: #007020; font-weight: bold } /* Keyword.Reserved */ .highlight .kt { color: #902000 } /* Keyword.Type */ .highlight .m { color: #208050 } /* Literal.Number */ -.highlight .s { color: #4070a0 } /* Literal.String */ -.highlight .na { color: #4070a0 } /* Name.Attribute */ +.highlight .s { color: #4070A0 } /* Literal.String */ +.highlight .na { color: #4070A0 } /* Name.Attribute */ .highlight .nb { color: #007020 } /* Name.Builtin */ -.highlight .nc { color: #0e84b5; font-weight: bold } /* Name.Class */ -.highlight .no { color: #60add5 } /* Name.Constant */ -.highlight .nd { color: #555555; font-weight: bold } /* Name.Decorator */ -.highlight .ni { color: #d55537; font-weight: bold } /* Name.Entity */ +.highlight .nc { color: #0E84B5; font-weight: bold } /* Name.Class */ +.highlight .no { color: #60ADD5 } /* Name.Constant */ +.highlight .nd { color: #555; font-weight: bold } /* Name.Decorator */ +.highlight .ni { color: #D55537; font-weight: bold } /* Name.Entity */ .highlight .ne { color: #007020 } /* Name.Exception */ -.highlight .nf { color: #06287e } /* Name.Function */ +.highlight .nf { color: #06287E } /* Name.Function */ .highlight .nl { color: #002070; font-weight: bold } /* Name.Label */ -.highlight .nn { color: #0e84b5; font-weight: bold } /* Name.Namespace */ +.highlight .nn { color: #0E84B5; font-weight: bold } /* Name.Namespace */ .highlight .nt { color: #062873; font-weight: bold } /* Name.Tag */ -.highlight .nv { color: #bb60d5 } /* Name.Variable */ +.highlight .nv { color: #BB60D5 } /* Name.Variable */ .highlight .ow { color: #007020; font-weight: bold } /* Operator.Word */ -.highlight .w { color: #bbbbbb } /* Text.Whitespace */ +.highlight .w { color: #BBB } /* Text.Whitespace */ .highlight .mb { color: #208050 } /* Literal.Number.Bin */ .highlight .mf { color: #208050 } /* Literal.Number.Float */ .highlight .mh { color: #208050 } /* Literal.Number.Hex */ .highlight .mi { color: #208050 } /* Literal.Number.Integer */ .highlight .mo { color: #208050 } /* Literal.Number.Oct */ -.highlight .sa { color: #4070a0 } /* Literal.String.Affix */ -.highlight .sb { color: #4070a0 } /* Literal.String.Backtick */ -.highlight .sc { color: #4070a0 } /* Literal.String.Char */ -.highlight .dl { color: #4070a0 } /* Literal.String.Delimiter */ -.highlight .sd { color: #4070a0; font-style: italic } /* Literal.String.Doc */ -.highlight .s2 { color: #4070a0 } /* Literal.String.Double */ -.highlight .se { color: #4070a0; font-weight: bold } /* Literal.String.Escape */ -.highlight .sh { color: #4070a0 } /* Literal.String.Heredoc */ -.highlight .si { color: #70a0d0; font-style: italic } /* Literal.String.Interpol */ -.highlight .sx { color: #c65d09 } /* Literal.String.Other */ +.highlight .sa { color: #4070A0 } /* Literal.String.Affix */ +.highlight .sb { color: #4070A0 } /* Literal.String.Backtick */ +.highlight .sc { color: #4070A0 } /* Literal.String.Char */ +.highlight .dl { color: #4070A0 } /* Literal.String.Delimiter */ +.highlight .sd { color: #4070A0; font-style: italic } /* Literal.String.Doc */ +.highlight .s2 { color: #4070A0 } /* Literal.String.Double */ +.highlight .se { color: #4070A0; font-weight: bold } /* Literal.String.Escape */ +.highlight .sh { color: #4070A0 } /* Literal.String.Heredoc */ +.highlight .si { color: #70A0D0; font-style: italic } /* Literal.String.Interpol */ +.highlight .sx { color: #C65D09 } /* Literal.String.Other */ .highlight .sr { color: #235388 } /* Literal.String.Regex */ -.highlight .s1 { color: #4070a0 } /* Literal.String.Single */ +.highlight .s1 { color: #4070A0 } /* Literal.String.Single */ .highlight .ss { color: #517918 } /* Literal.String.Symbol */ .highlight .bp { color: #007020 } /* Name.Builtin.Pseudo */ -.highlight .fm { color: #06287e } /* Name.Function.Magic */ -.highlight .vc { color: #bb60d5 } /* Name.Variable.Class */ -.highlight .vg { color: #bb60d5 } /* Name.Variable.Global */ -.highlight .vi { color: #bb60d5 } /* Name.Variable.Instance */ -.highlight .vm { color: #bb60d5 } /* Name.Variable.Magic */ +.highlight .fm { color: #06287E } /* Name.Function.Magic */ +.highlight .vc { color: #BB60D5 } /* Name.Variable.Class */ +.highlight .vg { color: #BB60D5 } /* Name.Variable.Global */ +.highlight .vi { color: #BB60D5 } /* Name.Variable.Instance */ +.highlight .vm { color: #BB60D5 } /* Name.Variable.Magic */ .highlight .il { color: #208050 } /* Literal.Number.Integer.Long */ \ No newline at end of file diff --git a/docs/cli/torchtrtc.html b/docs/cli/torchtrtc.html index 0332b3d7c4..0743462a93 100644 --- a/docs/cli/torchtrtc.html +++ b/docs/cli/torchtrtc.html @@ -10,7 +10,7 @@ - torchtrtc — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torchtrtc — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -851,7 +869,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -867,6 +885,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -905,6 +926,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -918,6 +942,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/contributors/conversion.html b/docs/contributors/conversion.html index 409ff24b05..7dc601418f 100644 --- a/docs/contributors/conversion.html +++ b/docs/contributors/conversion.html @@ -10,7 +10,7 @@ - Conversion Phase — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Conversion Phase — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -771,7 +789,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -787,6 +805,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -825,6 +846,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -838,6 +862,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/contributors/dynamo_converters.html b/docs/contributors/dynamo_converters.html index 3fa5662f82..54d746c989 100644 --- a/docs/contributors/dynamo_converters.html +++ b/docs/contributors/dynamo_converters.html @@ -10,7 +10,7 @@ - Writing Dynamo Converters — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Writing Dynamo Converters — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -480,7 +498,7 @@

    Converter implementation

    A converter is a function decrorated with torch_tensorrt.dynamo.dynamo_tensorrt_converter that follows the function signature:

    @torch_tensorrt.dynamo.conversion.dynamo_tensorrt_converter(torch.ops.aten.leaky_relu.default)
    -def leaky_relu_converter(
    +def leaky_relu_converter(
         ctx: torch_tensorrt.dynamo.conversion.ConversionCtx,
         target: Target,
         args: Tuple[Argument, ...],
    @@ -541,7 +559,7 @@ 

    Example: Convol 2: (np.ndarray, torch.Tensor, TRTTensor), } ) # type: ignore[misc] -def aten_ops_convolution( +def aten_ops_convolution( ctx: ConversionContext, target: Target, args: Tuple[Argument, ...], @@ -570,7 +588,7 @@

    Example: addmm<

    The decompositions are registered via register_torch_trt_decomposition decorator We define addmm_replacement and replace it with the torch ops, which will have their corresponding converters called.

    @torch_tensorrt.dynamo.lowering.register_torch_trt_decomposition(torch.ops.aten.addmm)
    -def addmm_replacement(
    +def addmm_replacement(
         input_: torch.Tensor, mat1: torch.Tensor, mat2: torch.Tensor, *, beta=1, alpha=1
     ) -> torch.Tensor:
         return torch.add(
    @@ -838,7 +856,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -854,6 +872,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -892,6 +913,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -905,6 +929,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/contributors/lowering.html b/docs/contributors/lowering.html index fb73bb1a46..beeb85f48f 100644 --- a/docs/contributors/lowering.html +++ b/docs/contributors/lowering.html @@ -10,7 +10,7 @@ - Lowering Phase — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Lowering Phase — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +

    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -941,7 +959,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -957,6 +975,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -995,6 +1016,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1008,6 +1032,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/contributors/partitioning.html b/docs/contributors/partitioning.html index 1750af6f79..d116fe8a86 100644 --- a/docs/contributors/partitioning.html +++ b/docs/contributors/partitioning.html @@ -10,7 +10,7 @@ - Partitioning Phase — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Partitioning Phase — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -938,7 +956,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -954,6 +972,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -992,6 +1013,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1005,6 +1029,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/contributors/phases.html b/docs/contributors/phases.html index dd31270654..c65216b1bf 100644 --- a/docs/contributors/phases.html +++ b/docs/contributors/phases.html @@ -10,7 +10,7 @@ - Compiler Phases — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Compiler Phases — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -739,7 +757,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -755,6 +773,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -793,6 +814,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -806,6 +830,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/contributors/runtime.html b/docs/contributors/runtime.html index 1fe3b7cee7..cf6cd73583 100644 --- a/docs/contributors/runtime.html +++ b/docs/contributors/runtime.html @@ -10,7 +10,7 @@ - Runtime Phase — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Runtime Phase — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -791,7 +809,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -807,6 +825,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -845,6 +866,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -858,6 +882,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/contributors/system_overview.html b/docs/contributors/system_overview.html index 48f2aa6653..80c52543fd 100644 --- a/docs/contributors/system_overview.html +++ b/docs/contributors/system_overview.html @@ -10,7 +10,7 @@ - System Overview — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + System Overview — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -775,7 +793,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -791,6 +809,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -829,6 +850,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -842,6 +866,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/contributors/ts_converters.html b/docs/contributors/ts_converters.html index f581ae2db4..8f8bb8beda 100644 --- a/docs/contributors/ts_converters.html +++ b/docs/contributors/ts_converters.html @@ -10,7 +10,7 @@ - Writing TorchScript Converters — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Writing TorchScript Converters — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -846,7 +864,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -862,6 +880,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -900,6 +921,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -913,6 +937,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/contributors/useful_links.html b/docs/contributors/useful_links.html index 0d153672ef..9bb792bff2 100644 --- a/docs/contributors/useful_links.html +++ b/docs/contributors/useful_links.html @@ -10,7 +10,7 @@ - Useful Links for Torch-TensorRT Development — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Useful Links for Torch-TensorRT Development — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -756,7 +774,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -772,6 +790,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -810,6 +831,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -823,6 +847,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/contributors/writing_dynamo_aten_lowering_passes.html b/docs/contributors/writing_dynamo_aten_lowering_passes.html index 5fa3939e9e..60b340a8f6 100644 --- a/docs/contributors/writing_dynamo_aten_lowering_passes.html +++ b/docs/contributors/writing_dynamo_aten_lowering_passes.html @@ -10,7 +10,7 @@ - Writing Dynamo ATen Lowering Passes — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Writing Dynamo ATen Lowering Passes — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -486,7 +504,7 @@

    Lowering Pass Requirements

    Example Lowering Pass

    -
    def repair_input_as_output(gm: torch.fx.GraphModule, sample_inputs: Sequence[torch.Tensor]) -> torch.fx.GraphModule:
    +
    diff --git a/docs/dynamo/dynamo_export.html b/docs/dynamo/dynamo_export.html
    index 4720fb9e54..aae33f9173 100644
    --- a/docs/dynamo/dynamo_export.html
    +++ b/docs/dynamo/dynamo_export.html
    @@ -10,7 +10,7 @@
     
       
       
    -  Compiling Exported Programs with Torch-TensorRT — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation
    +  Compiling Exported Programs with Torch-TensorRT — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation
       
     
       
    @@ -135,8 +135,8 @@
                       Developer Resources
                       

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +

    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -477,8 +495,8 @@ can export graphs from Pytorch programs into ExportedProgram objects. Torch-TensorRT dynamo frontend compiles these ExportedProgram objects and optimizes them using TensorRT. Here’s a simple usage of the dynamo frontend

    -
    import torch
    -import torch_tensorrt
    +
    import torch
    +import torch_tensorrt
     
     model = MyModel().eval().cuda()
     inputs = [torch.randn((1, 3, 224, 224), dtype=torch.float32).cuda()]
    @@ -525,8 +543,8 @@ 

    TracingExportedProgram can then be used with torch_tensorrt.dynamo.compile API. If you have dynamic input shapes in your model, you can use this torch_tensorrt.dynamo.trace to export the model with dynamic shapes. Alternatively, you can use torch.export with constraints directly as well.

    -
    import torch
    -import torch_tensorrt
    +
    import torch
    +import torch_tensorrt
     
     inputs = [torch_tensorrt.Input(min_shape=(1, 3, 224, 224),
                                   opt_shape=(4, 3, 224, 224),
    @@ -781,7 +799,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -797,6 +815,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -835,6 +856,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -848,6 +872,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/dynamo/torch_compile.html b/docs/dynamo/torch_compile.html index 868e9f175d..4942edf179 100644 --- a/docs/dynamo/torch_compile.html +++ b/docs/dynamo/torch_compile.html @@ -10,7 +10,7 @@ - TensorRT Backend for torch.compile — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + TensorRT Backend for torch.compile — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +

    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -477,7 +495,7 @@

    Key Features

    The primary goal of the Torch-TensorRT torch.compile backend is to enable Just-In-Time compilation workflows by combining the simplicity of torch.compile API with the performance of TensorRT. Invoking the torch.compile backend is as simple as importing the torch_tensorrt package and specifying the backend:

    -
    import torch_tensorrt
    +
    import torch_tensorrt
     ...
     optimized_model = torch.compile(model, backend="torch_tensorrt", dynamic=False)
     
    @@ -492,7 +510,7 @@

    Key Features

    -class torch_tensorrt.dynamo.CompilationSettings(enabled_precisions: ~typing.Set[~torch_tensorrt._enums.dtype] = <factory>, debug: bool = False, workspace_size: int = 0, min_block_size: int = 5, torch_executed_ops: ~typing.Collection[~typing.Union[~typing.Callable[[...], ~typing.Any], str]] = <factory>, pass_through_build_failures: bool = False, max_aux_streams: ~typing.Optional[int] = None, version_compatible: bool = False, optimization_level: ~typing.Optional[int] = None, use_python_runtime: ~typing.Optional[bool] = False, truncate_double: bool = False, use_fast_partitioner: bool = True, enable_experimental_decompositions: bool = False, device: ~torch_tensorrt._Device.Device = <factory>, require_full_compilation: bool = False, disable_tf32: bool = False, assume_dynamic_shape_support: bool = False, sparse_weights: bool = False, engine_capability: ~torch_tensorrt._enums.EngineCapability = <factory>, num_avg_timing_iters: int = 1, dla_sram_size: int = 1048576, dla_local_dram_size: int = 1073741824, dla_global_dram_size: int = 536870912, dryrun: ~typing.Union[bool, str] = False, hardware_compatible: bool = False, timing_cache_path: str = '/tmp/torch_tensorrt_engine_cache/timing_cache.bin', lazy_engine_init: bool = False, cache_built_engines: bool = False, reuse_cached_engines: bool = False, use_explicit_typing: bool = False, use_fp32_acc: bool = False, refit_identical_engine_weights: bool = False, strip_engine_weights: bool = False, immutable_weights: bool = True, enable_weight_streaming: bool = False, enable_cross_compile_for_windows: bool = False)[source]
    +class torch_tensorrt.dynamo.CompilationSettings(enabled_precisions: ~typing.Set[~torch_tensorrt._enums.dtype] = <factory>, debug: bool = False, workspace_size: int = 0, min_block_size: int = 5, torch_executed_ops: ~typing.Collection[~typing.Union[~typing.Callable[[...], ~typing.Any], str]] = <factory>, pass_through_build_failures: bool = False, max_aux_streams: ~typing.Optional[int] = None, version_compatible: bool = False, optimization_level: ~typing.Optional[int] = None, use_python_runtime: ~typing.Optional[bool] = False, truncate_double: bool = False, use_fast_partitioner: bool = True, enable_experimental_decompositions: bool = False, device: ~torch_tensorrt._Device.Device = <factory>, require_full_compilation: bool = False, disable_tf32: bool = False, assume_dynamic_shape_support: bool = False, sparse_weights: bool = False, engine_capability: ~torch_tensorrt._enums.EngineCapability = <factory>, num_avg_timing_iters: int = 1, dla_sram_size: int = 1048576, dla_local_dram_size: int = 1073741824, dla_global_dram_size: int = 536870912, dryrun: ~typing.Union[bool, str] = False, hardware_compatible: bool = False, timing_cache_path: str = '/tmp/torch_tensorrt_engine_cache/timing_cache.bin', lazy_engine_init: bool = False, cache_built_engines: bool = False, reuse_cached_engines: bool = False, use_explicit_typing: bool = False, use_fp32_acc: bool = False, refit_identical_engine_weights: bool = False, strip_engine_weights: bool = False, immutable_weights: bool = True, enable_weight_streaming: bool = False, enable_cross_compile_for_windows: bool = False, use_aot_joint_export: bool = True)[source]

    Compilation settings for Torch-TensorRT Dynamo Paths

    Parameters
    @@ -540,6 +558,7 @@

    Customizable Settings

    enable_weight_streaming (bool) – Enable weight streaming.

  • enable_cross_compile_for_windows (bool) – By default this is False means TensorRT engines can only be executed on the same platform where they were built. True will enable cross-platform compatibility which allows the engine to be built on Linux and run on Windows

  • +
  • use_aot_joint_export (bool) – Use aot_export_joint_simple, else wrap backend with AOT_autograd, required for distributed tensors

  • @@ -547,7 +566,7 @@

    Customizable Settings

    Custom Setting Usage

    -
    import torch_tensorrt
    +
    import torch_tensorrt
     ...
     optimized_model = torch.compile(model, backend="torch_tensorrt", dynamic=False,
                                     options={"truncate_long_and_double": True,
    @@ -568,7 +587,7 @@ 

    Custom Setting Usage

    Compilation

    Compilation is triggered by passing inputs to the model, as so:

    -
    import torch_tensorrt
    +
    import torch_tensorrt
     ...
     # Causes model compilation to occur
     first_outputs = optimized_model(*inputs)
    @@ -589,7 +608,7 @@ 

    Model Performance

    Operator Coverage

    Compilation is also a useful tool in determining operator coverage for a particular model. For instance, the following compilation command will display the operator coverage for each graph, but will not compile the model - effectively providing a “dryrun” mechanism:

    -
    import torch_tensorrt
    +
    import torch_tensorrt
     ...
     optimized_model = torch.compile(model, backend="torch_tensorrt", dynamic=False,
                                     options={"debug": True,
    @@ -601,8 +620,8 @@ 

    Operator Coverage

    Feasibility of Serialization

    Compilation can also be helpful in demonstrating graph breaks and the feasibility of serialization of a particular model. For instance, if a model has no graph breaks and compiles successfully with the Torch-TensorRT backend, then that model should be compilable and serializeable via the torch_tensorrt Dynamo IR, as discussed in Dynamic shapes with Torch-TensorRT. To determine the number of graph breaks in a model, the torch._dynamo.explain function is very useful:

    -
    import torch
    -import torch_tensorrt
    +
    import torch
    +import torch_tensorrt
     ...
     explanation = torch._dynamo.explain(model)(*inputs)
     print(f"Graph breaks: {explanation.graph_break_count}")
    @@ -875,7 +894,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -891,6 +910,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -929,6 +951,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -942,6 +967,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/fx/getting_started_with_fx_path.html b/docs/fx/getting_started_with_fx_path.html index d03f09d79d..73f4ec5262 100644 --- a/docs/fx/getting_started_with_fx_path.html +++ b/docs/fx/getting_started_with_fx_path.html @@ -10,7 +10,7 @@ - Torch-TensorRT (FX Frontend) User Guide — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Torch-TensorRT (FX Frontend) User Guide — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +

    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -1016,7 +1034,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1032,6 +1050,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1070,6 +1091,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1083,6 +1107,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/genindex.html b/docs/genindex.html index d032584364..98325928d9 100644 --- a/docs/genindex.html +++ b/docs/genindex.html @@ -9,7 +9,7 @@ - Index — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Index — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -132,8 +132,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -154,6 +154,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -201,6 +204,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -216,7 +223,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us

    @@ -272,7 +290,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -1427,7 +1445,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1443,6 +1461,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1481,6 +1502,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1494,6 +1518,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/getting_started/installation.html b/docs/getting_started/installation.html index fe15b213b1..ce8054a901 100644 --- a/docs/getting_started/installation.html +++ b/docs/getting_started/installation.html @@ -10,7 +10,7 @@ - Installation — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Installation — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +

    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -1115,7 +1133,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1131,6 +1149,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1169,6 +1190,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1182,6 +1206,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/getting_started/jetpack.html b/docs/getting_started/jetpack.html index b761bf3590..43fa62f4e7 100644 --- a/docs/getting_started/jetpack.html +++ b/docs/getting_started/jetpack.html @@ -10,7 +10,7 @@ - Overview — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Overview — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -811,7 +829,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -827,6 +845,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -865,6 +886,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -878,6 +902,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/getting_started/quick_start.html b/docs/getting_started/quick_start.html index 99d843281b..e37f815edf 100644 --- a/docs/getting_started/quick_start.html +++ b/docs/getting_started/quick_start.html @@ -10,7 +10,7 @@ - Quick Start — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Quick Start — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -476,8 +494,8 @@

    Option 1: torch.compile

    You can use Torch-TensorRT anywhere you use torch.compile:

    -
    import torch
    -import torch_tensorrt
    +
    import torch
    +import torch_tensorrt
     
     model = MyModel().eval().cuda() # define your model here
     x = torch.randn((1, 3, 224, 224)).cuda() # define what the inputs to the model will look like
    @@ -494,8 +512,8 @@ 

    Option 2: Export

    Step 1: Optimize + serialize

    -
    import torch
    -import torch_tensorrt
    +
    import torch
    +import torch_tensorrt
     
     model = MyModel().eval().cuda() # define your model here
     inputs = [torch.randn((1, 3, 224, 224)).cuda()] # define a list of representative inputs here
    @@ -510,8 +528,8 @@ 

    Step 1: Optimize + serialize

    Deployment in Python:

    -
    import torch
    -import torch_tensorrt
    +
    import torch
    +import torch_tensorrt
     
     inputs = [torch.randn((1, 3, 224, 224)).cuda()] # your inputs go here
     
    @@ -786,7 +804,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -802,6 +820,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -840,6 +861,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -853,6 +877,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/index.html b/docs/index.html index 3abe58676e..0807b3c505 100644 --- a/docs/index.html +++ b/docs/index.html @@ -10,7 +10,7 @@ - Torch-TensorRT — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Torch-TensorRT — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -134,8 +134,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -156,6 +156,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -203,6 +206,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -218,7 +225,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -274,7 +292,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -886,7 +904,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -902,6 +920,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -940,6 +961,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -953,6 +977,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/indices/supported_ops.html b/docs/indices/supported_ops.html index 4cac58a36f..77fe9ffc0e 100644 --- a/docs/indices/supported_ops.html +++ b/docs/indices/supported_ops.html @@ -10,7 +10,7 @@ - Operators Supported — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Operators Supported — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -134,8 +134,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -156,6 +156,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -203,6 +206,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -218,7 +225,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -274,7 +292,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -1002,7 +1020,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1018,6 +1036,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1056,6 +1077,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1069,6 +1093,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/objects.inv b/docs/objects.inv index a846d7f9d7..c32270832d 100644 Binary files a/docs/objects.inv and b/docs/objects.inv differ diff --git a/docs/py-modindex.html b/docs/py-modindex.html index 7b38f1cc29..f931caf1af 100644 --- a/docs/py-modindex.html +++ b/docs/py-modindex.html @@ -9,7 +9,7 @@ - Python Module Index — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Python Module Index — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -741,7 +759,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -757,6 +775,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -795,6 +816,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -808,6 +832,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/py_api/dynamo.html b/docs/py_api/dynamo.html index 5203926b74..d55fae495e 100644 --- a/docs/py_api/dynamo.html +++ b/docs/py_api/dynamo.html @@ -10,7 +10,7 @@ - torch_tensorrt.dynamo — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt.dynamo — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -663,7 +681,7 @@

    Functions

    -class torch_tensorrt.dynamo.CompilationSettings(enabled_precisions: ~typing.Set[~torch_tensorrt._enums.dtype] = <factory>, debug: bool = False, workspace_size: int = 0, min_block_size: int = 5, torch_executed_ops: ~typing.Collection[~typing.Union[~typing.Callable[[...], ~typing.Any], str]] = <factory>, pass_through_build_failures: bool = False, max_aux_streams: ~typing.Optional[int] = None, version_compatible: bool = False, optimization_level: ~typing.Optional[int] = None, use_python_runtime: ~typing.Optional[bool] = False, truncate_double: bool = False, use_fast_partitioner: bool = True, enable_experimental_decompositions: bool = False, device: ~torch_tensorrt._Device.Device = <factory>, require_full_compilation: bool = False, disable_tf32: bool = False, assume_dynamic_shape_support: bool = False, sparse_weights: bool = False, engine_capability: ~torch_tensorrt._enums.EngineCapability = <factory>, num_avg_timing_iters: int = 1, dla_sram_size: int = 1048576, dla_local_dram_size: int = 1073741824, dla_global_dram_size: int = 536870912, dryrun: ~typing.Union[bool, str] = False, hardware_compatible: bool = False, timing_cache_path: str = '/tmp/torch_tensorrt_engine_cache/timing_cache.bin', lazy_engine_init: bool = False, cache_built_engines: bool = False, reuse_cached_engines: bool = False, use_explicit_typing: bool = False, use_fp32_acc: bool = False, refit_identical_engine_weights: bool = False, strip_engine_weights: bool = False, immutable_weights: bool = True, enable_weight_streaming: bool = False, enable_cross_compile_for_windows: bool = False)[source]
    +class torch_tensorrt.dynamo.CompilationSettings(enabled_precisions: ~typing.Set[~torch_tensorrt._enums.dtype] = <factory>, debug: bool = False, workspace_size: int = 0, min_block_size: int = 5, torch_executed_ops: ~typing.Collection[~typing.Union[~typing.Callable[[...], ~typing.Any], str]] = <factory>, pass_through_build_failures: bool = False, max_aux_streams: ~typing.Optional[int] = None, version_compatible: bool = False, optimization_level: ~typing.Optional[int] = None, use_python_runtime: ~typing.Optional[bool] = False, truncate_double: bool = False, use_fast_partitioner: bool = True, enable_experimental_decompositions: bool = False, device: ~torch_tensorrt._Device.Device = <factory>, require_full_compilation: bool = False, disable_tf32: bool = False, assume_dynamic_shape_support: bool = False, sparse_weights: bool = False, engine_capability: ~torch_tensorrt._enums.EngineCapability = <factory>, num_avg_timing_iters: int = 1, dla_sram_size: int = 1048576, dla_local_dram_size: int = 1073741824, dla_global_dram_size: int = 536870912, dryrun: ~typing.Union[bool, str] = False, hardware_compatible: bool = False, timing_cache_path: str = '/tmp/torch_tensorrt_engine_cache/timing_cache.bin', lazy_engine_init: bool = False, cache_built_engines: bool = False, reuse_cached_engines: bool = False, use_explicit_typing: bool = False, use_fp32_acc: bool = False, refit_identical_engine_weights: bool = False, strip_engine_weights: bool = False, immutable_weights: bool = True, enable_weight_streaming: bool = False, enable_cross_compile_for_windows: bool = False, use_aot_joint_export: bool = True)[source]

    Compilation settings for Torch-TensorRT Dynamo Paths

    Parameters
    @@ -711,6 +729,7 @@

    ClassesDeveloper Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -976,6 +995,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1014,6 +1036,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1027,6 +1052,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/py_api/fx.html b/docs/py_api/fx.html index 2c952515d4..dea5e3d4c7 100644 --- a/docs/py_api/fx.html +++ b/docs/py_api/fx.html @@ -10,7 +10,7 @@ - torch_tensorrt.fx — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt.fx — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -789,7 +807,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -805,6 +823,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -843,6 +864,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -856,6 +880,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/py_api/logging.html b/docs/py_api/logging.html index 32d06a1517..4a2d9ec8bb 100644 --- a/docs/py_api/logging.html +++ b/docs/py_api/logging.html @@ -10,7 +10,7 @@ - torch_tensorrt.logging — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt.logging — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -781,7 +799,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -797,6 +815,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -835,6 +856,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -848,6 +872,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/py_api/ptq.html b/docs/py_api/ptq.html index 7ac17fb125..be608c6aa9 100644 --- a/docs/py_api/ptq.html +++ b/docs/py_api/ptq.html @@ -10,7 +10,7 @@ - torch_tensorrt.ts.ptq — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt.ts.ptq — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -783,7 +801,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -799,6 +817,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -837,6 +858,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -850,6 +874,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/py_api/runtime.html b/docs/py_api/runtime.html index f10ecaa1fe..15eb6502f5 100644 --- a/docs/py_api/runtime.html +++ b/docs/py_api/runtime.html @@ -10,7 +10,7 @@ - torch_tensorrt.runtime — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt.runtime — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -581,13 +599,13 @@

    Classes
    -class torch_tensorrt.runtime.PythonTorchTensorRTModule(serialized_engine: ~typing.Optional[bytes] = None, input_binding_names: ~typing.Optional[~typing.List[str]] = None, output_binding_names: ~typing.Optional[~typing.List[str]] = None, *, name: str = '', settings: ~torch_tensorrt.dynamo._settings.CompilationSettings = CompilationSettings(enabled_precisions={<dtype.f32: 7>}, debug=False, workspace_size=0, min_block_size=5, torch_executed_ops=set(), pass_through_build_failures=False, max_aux_streams=None, version_compatible=False, optimization_level=None, use_python_runtime=False, truncate_double=False, use_fast_partitioner=True, enable_experimental_decompositions=False, device=Device(type=DeviceType.GPU, gpu_id=0), require_full_compilation=False, disable_tf32=False, assume_dynamic_shape_support=False, sparse_weights=False, engine_capability=<EngineCapability.STANDARD: 1>, num_avg_timing_iters=1, dla_sram_size=1048576, dla_local_dram_size=1073741824, dla_global_dram_size=536870912, dryrun=False, hardware_compatible=False, timing_cache_path='/tmp/torch_tensorrt_engine_cache/timing_cache.bin', lazy_engine_init=False, cache_built_engines=False, reuse_cached_engines=False, use_explicit_typing=False, use_fp32_acc=False, refit_identical_engine_weights=False, strip_engine_weights=False, immutable_weights=True, enable_weight_streaming=False, enable_cross_compile_for_windows=False), weight_name_map: ~typing.Optional[dict[typing.Any, typing.Any]] = None)[source]
    +class torch_tensorrt.runtime.PythonTorchTensorRTModule(serialized_engine: ~typing.Optional[bytes] = None, input_binding_names: ~typing.Optional[~typing.List[str]] = None, output_binding_names: ~typing.Optional[~typing.List[str]] = None, *, name: str = '', settings: ~torch_tensorrt.dynamo._settings.CompilationSettings = CompilationSettings(enabled_precisions={<dtype.f32: 7>}, debug=False, workspace_size=0, min_block_size=5, torch_executed_ops=set(), pass_through_build_failures=False, max_aux_streams=None, version_compatible=False, optimization_level=None, use_python_runtime=False, truncate_double=False, use_fast_partitioner=True, enable_experimental_decompositions=False, device=Device(type=DeviceType.GPU, gpu_id=0), require_full_compilation=False, disable_tf32=False, assume_dynamic_shape_support=False, sparse_weights=False, engine_capability=<EngineCapability.STANDARD: 1>, num_avg_timing_iters=1, dla_sram_size=1048576, dla_local_dram_size=1073741824, dla_global_dram_size=536870912, dryrun=False, hardware_compatible=False, timing_cache_path='/tmp/torch_tensorrt_engine_cache/timing_cache.bin', lazy_engine_init=False, cache_built_engines=False, reuse_cached_engines=False, use_explicit_typing=False, use_fp32_acc=False, refit_identical_engine_weights=False, strip_engine_weights=False, immutable_weights=True, enable_weight_streaming=False, enable_cross_compile_for_windows=False, use_aot_joint_export=True), weight_name_map: ~typing.Optional[dict[typing.Any, typing.Any]] = None)[source]

    PythonTorchTensorRTModule is a PyTorch module which encompasses an arbitrary TensorRT Engine.

    This module is backed by the Torch-TensorRT runtime and is only compatible with FX / Dynamo / Python deployments. This module cannot be serialized to torchscript via torch.jit.trace for C++ deployment.

    -__init__(serialized_engine: ~typing.Optional[bytes] = None, input_binding_names: ~typing.Optional[~typing.List[str]] = None, output_binding_names: ~typing.Optional[~typing.List[str]] = None, *, name: str = '', settings: ~torch_tensorrt.dynamo._settings.CompilationSettings = CompilationSettings(enabled_precisions={<dtype.f32: 7>}, debug=False, workspace_size=0, min_block_size=5, torch_executed_ops=set(), pass_through_build_failures=False, max_aux_streams=None, version_compatible=False, optimization_level=None, use_python_runtime=False, truncate_double=False, use_fast_partitioner=True, enable_experimental_decompositions=False, device=Device(type=DeviceType.GPU, gpu_id=0), require_full_compilation=False, disable_tf32=False, assume_dynamic_shape_support=False, sparse_weights=False, engine_capability=<EngineCapability.STANDARD: 1>, num_avg_timing_iters=1, dla_sram_size=1048576, dla_local_dram_size=1073741824, dla_global_dram_size=536870912, dryrun=False, hardware_compatible=False, timing_cache_path='/tmp/torch_tensorrt_engine_cache/timing_cache.bin', lazy_engine_init=False, cache_built_engines=False, reuse_cached_engines=False, use_explicit_typing=False, use_fp32_acc=False, refit_identical_engine_weights=False, strip_engine_weights=False, immutable_weights=True, enable_weight_streaming=False, enable_cross_compile_for_windows=False), weight_name_map: ~typing.Optional[dict[typing.Any, typing.Any]] = None)[source]
    +__init__(serialized_engine: ~typing.Optional[bytes] = None, input_binding_names: ~typing.Optional[~typing.List[str]] = None, output_binding_names: ~typing.Optional[~typing.List[str]] = None, *, name: str = '', settings: ~torch_tensorrt.dynamo._settings.CompilationSettings = CompilationSettings(enabled_precisions={<dtype.f32: 7>}, debug=False, workspace_size=0, min_block_size=5, torch_executed_ops=set(), pass_through_build_failures=False, max_aux_streams=None, version_compatible=False, optimization_level=None, use_python_runtime=False, truncate_double=False, use_fast_partitioner=True, enable_experimental_decompositions=False, device=Device(type=DeviceType.GPU, gpu_id=0), require_full_compilation=False, disable_tf32=False, assume_dynamic_shape_support=False, sparse_weights=False, engine_capability=<EngineCapability.STANDARD: 1>, num_avg_timing_iters=1, dla_sram_size=1048576, dla_local_dram_size=1073741824, dla_global_dram_size=536870912, dryrun=False, hardware_compatible=False, timing_cache_path='/tmp/torch_tensorrt_engine_cache/timing_cache.bin', lazy_engine_init=False, cache_built_engines=False, reuse_cached_engines=False, use_explicit_typing=False, use_fp32_acc=False, refit_identical_engine_weights=False, strip_engine_weights=False, immutable_weights=True, enable_weight_streaming=False, enable_cross_compile_for_windows=False, use_aot_joint_export=True), weight_name_map: ~typing.Optional[dict[typing.Any, typing.Any]] = None)[source]

    Takes a name, target device, serialized TensorRT engine, and binding names / order and constructs a PyTorch torch.nn.Module around it. Uses TensorRT Python APIs to run the engine

    @@ -903,7 +921,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -919,6 +937,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -957,6 +978,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -970,6 +994,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/py_api/torch_tensorrt.html b/docs/py_api/torch_tensorrt.html index 1ebc412b8c..43503cbc18 100644 --- a/docs/py_api/torch_tensorrt.html +++ b/docs/py_api/torch_tensorrt.html @@ -10,7 +10,7 @@ - torch_tensorrt — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -1751,7 +1769,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1767,6 +1785,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1805,6 +1826,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1818,6 +1842,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/py_api/ts.html b/docs/py_api/ts.html index fe80f50f67..3956fa7518 100644 --- a/docs/py_api/ts.html +++ b/docs/py_api/ts.html @@ -10,7 +10,7 @@ - torch_tensorrt.ts — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + torch_tensorrt.ts — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -695,7 +713,7 @@

    Functions
    -torch_tensorrt.ts.TensorRTCompileSpec(inputs: Optional[List[torch.Tensor | Input]] = None, input_signature: Optional[Any] = None, device: Optional[Union[device, Device]] = None, disable_tf32: bool = False, sparse_weights: bool = False, enabled_precisions: Optional[Set[Union[dtype, dtype]]] = None, refit: bool = False, debug: bool = False, capability: EngineCapability = EngineCapability.STANDARD, num_avg_timing_iters: int = 1, workspace_size: int = 0, dla_sram_size: int = 1048576, dla_local_dram_size: int = 1073741824, dla_global_dram_size: int = 536870912, truncate_long_and_double: bool = False, calibrator: object = None, allow_shape_tensors: bool = False) <torch.ScriptClass object at 0x7f70888fd830>[source]
    +torch_tensorrt.ts.TensorRTCompileSpec(inputs: Optional[List[torch.Tensor | Input]] = None, input_signature: Optional[Any] = None, device: Optional[Union[device, Device]] = None, disable_tf32: bool = False, sparse_weights: bool = False, enabled_precisions: Optional[Set[Union[dtype, dtype]]] = None, refit: bool = False, debug: bool = False, capability: EngineCapability = EngineCapability.STANDARD, num_avg_timing_iters: int = 1, workspace_size: int = 0, dla_sram_size: int = 1048576, dla_local_dram_size: int = 1073741824, dla_global_dram_size: int = 536870912, truncate_long_and_double: bool = False, calibrator: object = None, allow_shape_tensors: bool = False) <torch.ScriptClass object at 0x7fe38a1f3730>[source]

    Utility to create a formatted spec dictionary for using the PyTorch TensorRT backend

    Keyword Arguments
    @@ -986,7 +1004,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1002,6 +1020,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1040,6 +1061,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1053,6 +1077,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/search.html b/docs/search.html index ab2599b327..642109fe1f 100644 --- a/docs/search.html +++ b/docs/search.html @@ -9,7 +9,7 @@ - Search — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Search — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -132,8 +132,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -154,6 +154,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -201,6 +204,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -216,7 +223,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -272,7 +290,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -709,7 +727,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -725,6 +743,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -763,6 +784,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -776,6 +800,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/searchindex.js b/docs/searchindex.js index 12ac2d9ebe..9c36cf3953 100644 --- a/docs/searchindex.js +++ b/docs/searchindex.js @@ -1 +1 @@ -Search.setIndex({"docnames": ["_cpp_api/classtorch__tensorrt_1_1DataType", "_cpp_api/classtorch__tensorrt_1_1Device_1_1DeviceType", "_cpp_api/classtorch__tensorrt_1_1TensorFormat", "_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8CacheCalibrator", "_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8Calibrator", "_cpp_api/define_macros_8h_1a18d295a837ac71add5578860b55e5502", "_cpp_api/define_macros_8h_1a282fd3c0b1c3a215148ae372070e1268", "_cpp_api/define_macros_8h_1a31398a6d4d27e28817afb0f0139e909e", "_cpp_api/define_macros_8h_1a35703561b26b1a9d2738ad7d58b27827", "_cpp_api/define_macros_8h_1abd1465eb38256d3f22cc1426b23d516b", "_cpp_api/define_macros_8h_1abe87b341f562fd1cf40b7672e4d759da", "_cpp_api/define_macros_8h_1ad19939408f7be171a74a89928b36eb59", "_cpp_api/define_macros_8h_1adad592a7b1b7eed529cdf6acd584c883", "_cpp_api/dir_cpp", "_cpp_api/dir_cpp_include", "_cpp_api/dir_cpp_include_torch_tensorrt", "_cpp_api/enum_logging_8h_1a130f65408ad8cbaee060f05e8db69558", "_cpp_api/enum_torch__tensorrt_8h_1a3fbe5d72e4fc624dbd038853079620eb", "_cpp_api/file_cpp_include_torch_tensorrt_logging.h", "_cpp_api/file_cpp_include_torch_tensorrt_macros.h", "_cpp_api/file_cpp_include_torch_tensorrt_ptq.h", "_cpp_api/file_cpp_include_torch_tensorrt_torch_tensorrt.h", "_cpp_api/function_logging_8h_1a0593f776f469c20469e2f729fc7861a3", "_cpp_api/function_logging_8h_1a0c012cb374addd90eb1f42eaec570650", "_cpp_api/function_logging_8h_1a56e110feaaba2c3fd44bd201fd21a76a", "_cpp_api/function_logging_8h_1a7cb50492421ea9de4e3db895819df6f2", "_cpp_api/function_logging_8h_1ac46ac0901cb97e3ae6e93b45f24e90b8", "_cpp_api/function_logging_8h_1ad2efd47b6c3689e58ccc595680579ae5", "_cpp_api/function_logging_8h_1af8f3443813315af7901903d25dd495cc", "_cpp_api/function_ptq_8h_1a226e3c83379d1012cde8578c1c86b16c", "_cpp_api/function_ptq_8h_1a6186e305f47c1d94b6130ef6c7f7e178", "_cpp_api/function_torch__tensorrt_8h_1a5b405fd3bf3c8fc2e2a54cbbab979797", "_cpp_api/function_torch__tensorrt_8h_1a6e19490a08fb1553c9dd347a5ae79db9", "_cpp_api/function_torch__tensorrt_8h_1a81f9783517335dda877d8cfcf38987c9", "_cpp_api/function_torch__tensorrt_8h_1ac4ab8313ae72c2c899ea31548b528528", "_cpp_api/function_torch__tensorrt_8h_1ad1acd06eaeaffbbcf6e7ebf426891384", "_cpp_api/function_torch__tensorrt_8h_1ad6a4ee8ca6c8f6e5519eb1128ec7f4a1", "_cpp_api/function_torch__tensorrt_8h_1ae8d56472106eeef37fbe51ff7f40c9b2", "_cpp_api/namespace_torch_tensorrt", "_cpp_api/namespace_torch_tensorrt__logging", "_cpp_api/namespace_torch_tensorrt__ptq", "_cpp_api/namespace_torch_tensorrt__torchscript", "_cpp_api/program_listing_file_cpp_include_torch_tensorrt_logging.h", "_cpp_api/program_listing_file_cpp_include_torch_tensorrt_macros.h", "_cpp_api/program_listing_file_cpp_include_torch_tensorrt_ptq.h", "_cpp_api/program_listing_file_cpp_include_torch_tensorrt_torch_tensorrt.h", "_cpp_api/structtorch__tensorrt_1_1Device", "_cpp_api/structtorch__tensorrt_1_1GraphInputs", "_cpp_api/structtorch__tensorrt_1_1Input", "_cpp_api/structtorch__tensorrt_1_1torchscript_1_1CompileSpec", "_cpp_api/torch_tensort_cpp", "_cpp_api/unabridged_orphan", "cli/torchtrtc", "contributors/conversion", "contributors/dynamo_converters", "contributors/lowering", "contributors/partitioning", "contributors/phases", "contributors/runtime", "contributors/system_overview", "contributors/ts_converters", "contributors/useful_links", "contributors/writing_dynamo_aten_lowering_passes", "dynamo/dynamo_export", "dynamo/torch_compile", "fx/getting_started_with_fx_path", "getting_started/installation", "getting_started/jetpack", "getting_started/quick_start", "index", "indices/supported_ops", "py_api/dynamo", "py_api/fx", "py_api/logging", "py_api/ptq", "py_api/runtime", "py_api/torch_tensorrt", "py_api/ts", "sg_execution_times", "src/pytorch-sphinx-theme/docs/changelog", "src/pytorch-sphinx-theme/docs/configuring", "src/pytorch-sphinx-theme/docs/demo/api", "src/pytorch-sphinx-theme/docs/demo/demo", "src/pytorch-sphinx-theme/docs/demo/lists_tables", "src/pytorch-sphinx-theme/docs/demo/long", "src/pytorch-sphinx-theme/docs/demo/structure", "src/pytorch-sphinx-theme/docs/index", "src/pytorch-sphinx-theme/docs/installing", "ts/creating_torchscript_module_in_python", "ts/getting_started_with_cpp_api", "ts/getting_started_with_python_api", "ts/ptq", "ts/torchscript_frontend_from_pytorch", "tutorials/_rendered_examples/dynamo/auto_generate_converters", "tutorials/_rendered_examples/dynamo/converter_overloading", "tutorials/_rendered_examples/dynamo/cross_runtime_compilation_for_windows", "tutorials/_rendered_examples/dynamo/custom_kernel_plugins", "tutorials/_rendered_examples/dynamo/engine_caching_bert_example", "tutorials/_rendered_examples/dynamo/engine_caching_example", "tutorials/_rendered_examples/dynamo/index", "tutorials/_rendered_examples/dynamo/mutable_torchtrt_module_example", "tutorials/_rendered_examples/dynamo/pre_allocated_output_example", "tutorials/_rendered_examples/dynamo/refit_engine_example", "tutorials/_rendered_examples/dynamo/torch_compile_advanced_usage", "tutorials/_rendered_examples/dynamo/torch_compile_gpt2", "tutorials/_rendered_examples/dynamo/torch_compile_resnet_example", "tutorials/_rendered_examples/dynamo/torch_compile_stable_diffusion", "tutorials/_rendered_examples/dynamo/torch_compile_transformers_example", "tutorials/_rendered_examples/dynamo/torch_export_cudagraphs", "tutorials/_rendered_examples/dynamo/torch_export_gpt2", "tutorials/_rendered_examples/dynamo/torch_export_llama2", "tutorials/_rendered_examples/dynamo/torch_export_sam2", "tutorials/_rendered_examples/dynamo/vgg16_ptq", "tutorials/_rendered_examples/dynamo/weight_streaming_example", "tutorials/_rendered_examples/index", "tutorials/_rendered_examples/triton/index", "tutorials/notebooks", "tutorials/serving_torch_tensorrt_with_triton", "user_guide/dynamic_shapes", "user_guide/mixed_precision", "user_guide/runtime", "user_guide/saving_models", "user_guide/torch_tensorrt_explained", "user_guide/using_dla"], "filenames": ["_cpp_api/classtorch__tensorrt_1_1DataType.rst", "_cpp_api/classtorch__tensorrt_1_1Device_1_1DeviceType.rst", "_cpp_api/classtorch__tensorrt_1_1TensorFormat.rst", "_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8CacheCalibrator.rst", "_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8Calibrator.rst", "_cpp_api/define_macros_8h_1a18d295a837ac71add5578860b55e5502.rst", "_cpp_api/define_macros_8h_1a282fd3c0b1c3a215148ae372070e1268.rst", "_cpp_api/define_macros_8h_1a31398a6d4d27e28817afb0f0139e909e.rst", "_cpp_api/define_macros_8h_1a35703561b26b1a9d2738ad7d58b27827.rst", "_cpp_api/define_macros_8h_1abd1465eb38256d3f22cc1426b23d516b.rst", "_cpp_api/define_macros_8h_1abe87b341f562fd1cf40b7672e4d759da.rst", "_cpp_api/define_macros_8h_1ad19939408f7be171a74a89928b36eb59.rst", "_cpp_api/define_macros_8h_1adad592a7b1b7eed529cdf6acd584c883.rst", "_cpp_api/dir_cpp.rst", "_cpp_api/dir_cpp_include.rst", "_cpp_api/dir_cpp_include_torch_tensorrt.rst", "_cpp_api/enum_logging_8h_1a130f65408ad8cbaee060f05e8db69558.rst", "_cpp_api/enum_torch__tensorrt_8h_1a3fbe5d72e4fc624dbd038853079620eb.rst", "_cpp_api/file_cpp_include_torch_tensorrt_logging.h.rst", "_cpp_api/file_cpp_include_torch_tensorrt_macros.h.rst", "_cpp_api/file_cpp_include_torch_tensorrt_ptq.h.rst", "_cpp_api/file_cpp_include_torch_tensorrt_torch_tensorrt.h.rst", "_cpp_api/function_logging_8h_1a0593f776f469c20469e2f729fc7861a3.rst", "_cpp_api/function_logging_8h_1a0c012cb374addd90eb1f42eaec570650.rst", "_cpp_api/function_logging_8h_1a56e110feaaba2c3fd44bd201fd21a76a.rst", "_cpp_api/function_logging_8h_1a7cb50492421ea9de4e3db895819df6f2.rst", "_cpp_api/function_logging_8h_1ac46ac0901cb97e3ae6e93b45f24e90b8.rst", "_cpp_api/function_logging_8h_1ad2efd47b6c3689e58ccc595680579ae5.rst", "_cpp_api/function_logging_8h_1af8f3443813315af7901903d25dd495cc.rst", "_cpp_api/function_ptq_8h_1a226e3c83379d1012cde8578c1c86b16c.rst", "_cpp_api/function_ptq_8h_1a6186e305f47c1d94b6130ef6c7f7e178.rst", "_cpp_api/function_torch__tensorrt_8h_1a5b405fd3bf3c8fc2e2a54cbbab979797.rst", "_cpp_api/function_torch__tensorrt_8h_1a6e19490a08fb1553c9dd347a5ae79db9.rst", "_cpp_api/function_torch__tensorrt_8h_1a81f9783517335dda877d8cfcf38987c9.rst", "_cpp_api/function_torch__tensorrt_8h_1ac4ab8313ae72c2c899ea31548b528528.rst", "_cpp_api/function_torch__tensorrt_8h_1ad1acd06eaeaffbbcf6e7ebf426891384.rst", "_cpp_api/function_torch__tensorrt_8h_1ad6a4ee8ca6c8f6e5519eb1128ec7f4a1.rst", "_cpp_api/function_torch__tensorrt_8h_1ae8d56472106eeef37fbe51ff7f40c9b2.rst", "_cpp_api/namespace_torch_tensorrt.rst", "_cpp_api/namespace_torch_tensorrt__logging.rst", "_cpp_api/namespace_torch_tensorrt__ptq.rst", "_cpp_api/namespace_torch_tensorrt__torchscript.rst", "_cpp_api/program_listing_file_cpp_include_torch_tensorrt_logging.h.rst", "_cpp_api/program_listing_file_cpp_include_torch_tensorrt_macros.h.rst", "_cpp_api/program_listing_file_cpp_include_torch_tensorrt_ptq.h.rst", "_cpp_api/program_listing_file_cpp_include_torch_tensorrt_torch_tensorrt.h.rst", "_cpp_api/structtorch__tensorrt_1_1Device.rst", "_cpp_api/structtorch__tensorrt_1_1GraphInputs.rst", "_cpp_api/structtorch__tensorrt_1_1Input.rst", "_cpp_api/structtorch__tensorrt_1_1torchscript_1_1CompileSpec.rst", "_cpp_api/torch_tensort_cpp.rst", "_cpp_api/unabridged_orphan.rst", "cli/torchtrtc.rst", "contributors/conversion.rst", "contributors/dynamo_converters.rst", "contributors/lowering.rst", "contributors/partitioning.rst", "contributors/phases.rst", "contributors/runtime.rst", "contributors/system_overview.rst", "contributors/ts_converters.rst", "contributors/useful_links.rst", "contributors/writing_dynamo_aten_lowering_passes.rst", "dynamo/dynamo_export.rst", "dynamo/torch_compile.rst", "fx/getting_started_with_fx_path.rst", "getting_started/installation.rst", "getting_started/jetpack.rst", "getting_started/quick_start.rst", "index.rst", "indices/supported_ops.rst", "py_api/dynamo.rst", "py_api/fx.rst", "py_api/logging.rst", "py_api/ptq.rst", "py_api/runtime.rst", "py_api/torch_tensorrt.rst", "py_api/ts.rst", "sg_execution_times.rst", "src/pytorch-sphinx-theme/docs/changelog.rst", "src/pytorch-sphinx-theme/docs/configuring.rst", "src/pytorch-sphinx-theme/docs/demo/api.rst", "src/pytorch-sphinx-theme/docs/demo/demo.rst", "src/pytorch-sphinx-theme/docs/demo/lists_tables.rst", "src/pytorch-sphinx-theme/docs/demo/long.rst", "src/pytorch-sphinx-theme/docs/demo/structure.rst", "src/pytorch-sphinx-theme/docs/index.rst", "src/pytorch-sphinx-theme/docs/installing.rst", "ts/creating_torchscript_module_in_python.rst", "ts/getting_started_with_cpp_api.rst", "ts/getting_started_with_python_api.rst", "ts/ptq.rst", "ts/torchscript_frontend_from_pytorch.rst", "tutorials/_rendered_examples/dynamo/auto_generate_converters.rst", "tutorials/_rendered_examples/dynamo/converter_overloading.rst", "tutorials/_rendered_examples/dynamo/cross_runtime_compilation_for_windows.rst", "tutorials/_rendered_examples/dynamo/custom_kernel_plugins.rst", "tutorials/_rendered_examples/dynamo/engine_caching_bert_example.rst", "tutorials/_rendered_examples/dynamo/engine_caching_example.rst", "tutorials/_rendered_examples/dynamo/index.rst", "tutorials/_rendered_examples/dynamo/mutable_torchtrt_module_example.rst", "tutorials/_rendered_examples/dynamo/pre_allocated_output_example.rst", "tutorials/_rendered_examples/dynamo/refit_engine_example.rst", "tutorials/_rendered_examples/dynamo/torch_compile_advanced_usage.rst", "tutorials/_rendered_examples/dynamo/torch_compile_gpt2.rst", "tutorials/_rendered_examples/dynamo/torch_compile_resnet_example.rst", "tutorials/_rendered_examples/dynamo/torch_compile_stable_diffusion.rst", "tutorials/_rendered_examples/dynamo/torch_compile_transformers_example.rst", "tutorials/_rendered_examples/dynamo/torch_export_cudagraphs.rst", "tutorials/_rendered_examples/dynamo/torch_export_gpt2.rst", "tutorials/_rendered_examples/dynamo/torch_export_llama2.rst", "tutorials/_rendered_examples/dynamo/torch_export_sam2.rst", "tutorials/_rendered_examples/dynamo/vgg16_ptq.rst", "tutorials/_rendered_examples/dynamo/weight_streaming_example.rst", "tutorials/_rendered_examples/index.rst", "tutorials/_rendered_examples/triton/index.rst", "tutorials/notebooks.rst", "tutorials/serving_torch_tensorrt_with_triton.rst", "user_guide/dynamic_shapes.rst", "user_guide/mixed_precision.rst", "user_guide/runtime.rst", "user_guide/saving_models.rst", "user_guide/torch_tensorrt_explained.rst", "user_guide/using_dla.rst"], "titles": ["Class DataType", "Class Device::DeviceType", "Class TensorFormat", "Template Class Int8CacheCalibrator", "Template Class Int8Calibrator", "Define STR", "Define TORCH_TENSORRT_PATCH_VERSION", "Define TORCH_TENSORRT_MAJOR_VERSION", "Define TORCH_TENSORRT_MINOR_VERSION", "Define TORCHTRT_API", "Define XSTR", "Define TORCHTRT_HIDDEN", "Define TORCH_TENSORRT_VERSION", "Directory cpp", "Directory include", "Directory torch_tensorrt", "Enum Level", "Enum EngineCapability", "File logging.h", "File macros.h", "File ptq.h", "File torch_tensorrt.h", "Function torch_tensorrt::logging::get_logging_prefix", "Function torch_tensorrt::logging::get_reportable_log_level", "Function torch_tensorrt::logging::get_is_colored_output_on", "Function torch_tensorrt::logging::set_reportable_log_level", "Function torch_tensorrt::logging::log", "Function torch_tensorrt::logging::set_is_colored_output_on", "Function torch_tensorrt::logging::set_logging_prefix", "Template Function torch_tensorrt::ptq::make_int8_cache_calibrator", "Template Function torch_tensorrt::ptq::make_int8_calibrator", "Function torch_tensorrt::torchscript::check_method_operator_support", "Function torch_tensorrt::torchscript::compile", "Function torch_tensorrt::torchscript::embed_engine_in_new_module", "Function torch_tensorrt::get_build_info", "Function torch_tensorrt::set_device", "Function torch_tensorrt::dump_build_info", "Function torch_tensorrt::torchscript::convert_method_to_trt_engine", "Namespace torch_tensorrt", "Namespace torch_tensorrt::logging", "Namespace torch_tensorrt::ptq", "Namespace torch_tensorrt::torchscript", "Program Listing for File logging.h", "Program Listing for File macros.h", "Program Listing for File ptq.h", "Program Listing for File torch_tensorrt.h", "Struct Device", "Struct GraphInputs", "Struct Input", "Struct CompileSpec", "Torch-TensorRT C++ API", "Full API", "torchtrtc", "Conversion Phase", "Writing Dynamo Converters", "Lowering Phase", "Partitioning Phase", "Compiler Phases", "Runtime Phase", "System Overview", "Writing TorchScript Converters", "Useful Links for Torch-TensorRT Development", "Writing Dynamo ATen Lowering Passes", "Compiling Exported Programs with Torch-TensorRT", "TensorRT Backend for torch.compile", "Torch-TensorRT (FX Frontend) User Guide", "Installation", "Overview", "Quick Start", "Torch-TensorRT", "Operators Supported", "torch_tensorrt.dynamo", "torch_tensorrt.fx", "torch_tensorrt.logging", "torch_tensorrt.ts.ptq", "torch_tensorrt.runtime", "torch_tensorrt", "torch_tensorrt.ts", "Computation times", "Changelog", "Configuration", "5. :mod:`test_py_module`", "3. Paragraph Level Markup", "4. Lists & Tables", "1. Long Sticky Nav", "1. Structural Elements", "<no title>", "Installation", "Creating a TorchScript Module", "Using Torch-TensorRT in C++", "Using Torch-TensorRT in Python", "Post Training Quantization (PTQ)", "Using Torch-TensorRT TorchScript Frontend Directly From PyTorch", "Automatically Generate a Converter for a Custom Kernel", "Overloading Torch-TensorRT Converters with Custom Converters", "Cross runtime compilation for windows example", "Using Custom Kernels within TensorRT Engines with Torch-TensorRT", "Engine Caching (BERT)", "Engine Caching", "Dependencies", "Mutable Torch TensorRT Module", "Pre-allocated output buffer", "Refitting Torch-TensorRT Programs with New Weights", "Torch Compile Advanced Usage", "Compiling GPT2 using the Torch-TensorRT torch.compile frontend", "Compiling ResNet with dynamic shapes using the torch.compile backend", "Compiling Stable Diffusion model using the torch.compile backend", "Compiling BERT using the torch.compile backend", "Torch Export with Cudagraphs", "Compiling GPT2 using the dynamo backend", "Compiling Llama2 using the dynamo backend", "Compiling SAM2 using the dynamo backend", "Deploy Quantized Models using Torch-TensorRT", "Weight Streaming", "Torch-TensorRT Tutorials", "Serving a Torch-TensorRT model with Triton", "Legacy notebooks", "Serving a Torch-TensorRT model with Triton", "Dynamic shapes with Torch-TensorRT", "Compile Mixed Precision models with Torch-TensorRT", "Deploying Torch-TensorRT Programs", "Saving models compiled with Torch-TensorRT", "Torch-TensorRT Explained", "DLA"], "terms": {"defin": [0, 1, 2, 3, 4, 16, 17, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 43, 46, 47, 48, 49, 51, 52, 54, 65, 68, 75, 76, 80, 88, 89, 90, 91, 93, 94, 96, 98, 103, 107, 108, 109, 110, 116], "file": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 16, 17, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 46, 47, 48, 49, 52, 54, 56, 58, 59, 64, 65, 66, 67, 68, 71, 72, 74, 76, 77, 78, 80, 81, 83, 87, 89, 91, 95, 114, 115, 117, 118, 121], "torch_tensorrt": [0, 1, 2, 14, 16, 17, 42, 43, 44, 46, 47, 48, 49, 50, 51, 52, 54, 56, 62, 63, 64, 65, 68, 69, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 106, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 123], "h": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 15, 16, 17, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 46, 47, 48, 49, 50, 51, 52, 55, 68, 76, 89, 91, 111], "support": [0, 1, 2, 27, 31, 46, 48, 49, 52, 54, 56, 61, 63, 65, 67, 68, 69, 72, 75, 76, 77, 80, 81, 88, 89, 90, 93, 94, 96, 101, 102, 104, 105, 107, 109, 110, 112, 113, 114, 115, 117, 119, 122, 123], "data": [0, 2, 3, 4, 29, 30, 44, 46, 48, 49, 52, 53, 56, 57, 59, 60, 64, 65, 70, 71, 72, 74, 76, 77, 82, 86, 90, 91, 93, 96, 98, 104, 111, 112, 113, 116], "type": [0, 1, 2, 30, 49, 50, 52, 53, 56, 58, 60, 62, 63, 64, 65, 71, 72, 74, 75, 76, 77, 82, 89, 90, 91, 93, 94, 95, 96, 98, 111, 112, 113, 116, 119, 121], "can": [0, 1, 4, 29, 30, 37, 46, 47, 48, 49, 52, 53, 54, 55, 56, 57, 58, 59, 60, 62, 63, 64, 65, 66, 67, 68, 71, 74, 75, 76, 77, 80, 82, 88, 89, 90, 91, 92, 93, 94, 95, 96, 98, 100, 101, 102, 103, 104, 107, 108, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122], "us": [0, 1, 2, 3, 4, 29, 30, 32, 35, 37, 43, 44, 45, 46, 48, 49, 52, 53, 54, 56, 58, 59, 60, 62, 63, 64, 65, 66, 67, 68, 69, 71, 72, 74, 75, 76, 77, 78, 80, 81, 82, 83, 88, 91, 95, 98, 99, 100, 102, 113, 114, 115, 117, 119, 120, 121, 122, 123], "tensorrt": [0, 1, 3, 4, 29, 30, 31, 32, 33, 36, 37, 44, 45, 46, 48, 49, 52, 53, 54, 55, 56, 57, 59, 60, 62, 67, 68, 71, 72, 74, 75, 76, 77, 88, 91, 95, 98, 99, 101, 103, 105, 106, 107, 108, 113], "engin": [0, 1, 17, 32, 33, 37, 45, 46, 48, 49, 52, 53, 56, 57, 59, 62, 63, 64, 69, 71, 72, 75, 76, 77, 80, 89, 90, 91, 92, 93, 94, 99, 101, 102, 104, 105, 107, 113, 114, 118, 120, 122, 123], "thi": [0, 1, 2, 29, 30, 42, 43, 44, 45, 46, 47, 48, 49, 52, 53, 54, 55, 56, 57, 58, 59, 60, 62, 63, 64, 65, 66, 67, 68, 71, 72, 75, 76, 77, 80, 81, 82, 84, 85, 88, 89, 91, 92, 93, 94, 96, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122], "compat": [0, 1, 46, 55, 58, 64, 65, 71, 75, 76, 77, 111, 122], "c10": [0, 1, 45, 46, 48, 49, 89, 91], "check": [0, 1, 31, 46, 52, 55, 60, 65, 67, 71, 75, 77, 89, 96, 100, 102, 114, 115, 117, 120], "trt": [0, 1, 3, 4, 46, 48, 53, 55, 58, 60, 62, 64, 65, 67, 68, 70, 71, 75, 76, 89, 94, 96, 101, 104, 107, 109, 110, 111, 113, 118, 120, 121], "so": [0, 44, 52, 53, 54, 55, 58, 59, 60, 62, 64, 65, 66, 67, 72, 75, 76, 81, 82, 83, 89, 91, 93, 94, 96, 98, 103, 104, 105, 107, 109, 110, 118], "should": [0, 3, 4, 29, 45, 49, 52, 53, 54, 55, 56, 57, 59, 60, 63, 64, 65, 67, 71, 75, 76, 77, 80, 82, 85, 91, 94, 96, 97, 98, 101, 102, 104, 108, 111, 114, 115, 117], "reason": [0, 65, 88, 94, 96, 98, 122], "you": [0, 1, 2, 29, 30, 46, 48, 49, 52, 53, 54, 55, 56, 58, 59, 60, 63, 65, 66, 67, 68, 71, 75, 76, 77, 80, 82, 83, 84, 88, 89, 90, 91, 92, 93, 94, 96, 98, 99, 100, 102, 108, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122], "need": [0, 1, 2, 25, 29, 43, 46, 53, 54, 55, 60, 65, 66, 67, 71, 72, 75, 76, 82, 89, 90, 91, 93, 94, 96, 97, 98, 100, 102, 114, 115, 116, 117, 118, 120], "explictli": 0, "public": [0, 1, 2, 3, 4, 44, 45, 46, 47, 48, 49, 83, 91], "enum": [0, 1, 2, 42, 45, 46, 51, 71, 77, 91, 94], "valu": [0, 1, 2, 16, 17, 45, 46, 48, 53, 56, 58, 60, 63, 70, 71, 74, 76, 80, 89, 100, 103, 104, 105, 107, 113, 116], "underli": [0, 1, 2, 46, 60], "In": [0, 1, 2, 46, 53, 54, 56, 57, 58, 59, 60, 64, 65, 66, 75, 76, 82, 83, 85, 90, 91, 93, 94, 96, 100, 104, 111, 114, 115, 116, 117, 118, 119, 120, 121], "case": [0, 1, 2, 46, 49, 53, 54, 56, 58, 60, 62, 64, 65, 66, 67, 75, 76, 91, 93, 94, 96, 100, 101, 102, 118, 119, 120], "itself": [0, 1, 2, 46, 52, 55, 92, 94, 114, 115, 117], "interfac": [0, 1, 2, 46, 58, 59, 60, 64, 69, 91], "vs": [0, 1, 2, 46, 55, 66, 71, 76, 77, 92], "normal": [0, 1, 2, 46, 65, 82, 88, 89, 91, 94, 100, 101, 102, 108, 111, 112, 114, 115, 117, 123], "instatin": [0, 1, 2, 46], "ex": [0, 1, 2, 33, 46, 67, 77, 83, 85], "kfloat": [0, 45, 49], "enumer": [0, 1, 2, 16, 17, 46, 111], "klong": [0, 45], "int64": [0, 76, 77, 113], "kdoubl": [0, 45], "fp64": [0, 76], "fp32": [0, 48, 49, 52, 64, 65, 71, 76, 77, 91, 109, 110, 111, 114, 115, 116, 117, 119], "khalf": [0, 45, 89], "fp16": [0, 48, 49, 52, 64, 65, 71, 72, 76, 89, 90, 100, 106, 109, 110, 111, 113, 119, 123], "kchar": [0, 45], "int8": [0, 44, 48, 49, 52, 64, 71, 76, 77, 91, 112, 123], "kint": [0, 45], "int": [0, 3, 4, 35, 44, 45, 49, 52, 54, 56, 63, 64, 70, 71, 72, 76, 77, 80, 89, 93, 96, 111, 112, 113], "kbool": [0, 45], "bool": [0, 1, 2, 3, 4, 24, 27, 30, 31, 42, 44, 45, 46, 49, 55, 60, 64, 70, 71, 72, 74, 75, 76, 77, 80, 89, 91, 95, 96], "kunknown": [0, 2, 45], "sentinel": [0, 2, 76], "function": [0, 1, 2, 3, 4, 46, 48, 49, 51, 54, 55, 56, 58, 60, 62, 64, 65, 66, 88, 89, 91, 92, 93, 94, 96, 102, 103, 107, 108, 109, 110, 111, 114, 115, 116, 117, 118, 120, 122, 123], "default": [0, 1, 2, 3, 4, 16, 29, 30, 33, 43, 45, 46, 48, 49, 52, 54, 56, 62, 64, 65, 66, 71, 72, 75, 76, 77, 80, 81, 82, 89, 90, 91, 92, 93, 94, 95, 96, 98, 112, 118, 120, 121, 122], "construct": [0, 1, 2, 3, 4, 46, 48, 49, 53, 54, 55, 57, 59, 60, 65, 74, 75, 76, 82, 83, 89, 91, 94, 96, 98, 118], "new": [0, 1, 2, 3, 4, 32, 33, 46, 48, 49, 56, 58, 59, 60, 62, 64, 65, 68, 69, 71, 77, 82, 89, 93, 98, 99, 100, 101, 104, 105, 107, 108, 114, 115, 117, 120], "object": [0, 1, 2, 3, 4, 46, 48, 49, 52, 58, 60, 62, 63, 64, 71, 75, 76, 77, 91, 92, 94, 101, 118, 121], "inlin": [0, 1, 2, 3, 4, 29, 30, 44, 46, 48, 55, 83, 86, 89], "constexpr": [0, 1, 2, 45, 46, 93, 96], "t": [0, 1, 2, 45, 46, 55, 60, 65, 66, 70, 76, 80, 82, 83, 88, 89, 91, 93, 94, 96, 112, 114, 115, 117, 118], "constructor": [0, 2, 46, 48, 49, 58, 88], "from": [0, 1, 2, 3, 4, 29, 30, 44, 46, 48, 49, 52, 53, 55, 56, 57, 58, 59, 60, 63, 64, 65, 67, 69, 71, 72, 75, 76, 77, 78, 80, 81, 82, 83, 88, 89, 91, 93, 94, 95, 96, 97, 98, 100, 101, 102, 104, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 120, 121, 122], "torchtrt_api": [0, 2, 19, 22, 23, 24, 25, 26, 27, 28, 31, 32, 33, 34, 35, 36, 37, 42, 43, 44, 45, 48, 49, 50], "scalartyp": [0, 45, 70], "torch": [0, 1, 2, 4, 20, 21, 29, 30, 31, 32, 33, 36, 37, 44, 45, 46, 47, 48, 49, 52, 53, 54, 55, 56, 57, 58, 59, 60, 62, 67, 71, 72, 74, 75, 76, 77, 78, 88, 91, 93, 95, 97, 98, 99, 101, 113, 123], "paramet": [0, 1, 2, 3, 4, 25, 26, 27, 29, 30, 31, 32, 33, 35, 37, 46, 48, 49, 53, 54, 55, 60, 64, 65, 71, 72, 74, 75, 76, 77, 86, 88, 89, 102, 109, 110], "oper": [0, 1, 2, 3, 4, 31, 44, 45, 46, 49, 52, 53, 55, 56, 57, 58, 59, 60, 62, 63, 65, 69, 71, 76, 77, 90, 91, 94, 101, 102, 105, 107, 108, 111, 122, 123], "const": [0, 1, 2, 3, 4, 29, 30, 31, 32, 33, 35, 37, 44, 45, 46, 55, 60, 70, 89, 91], "get": [0, 1, 2, 3, 4, 23, 34, 44, 46, 55, 56, 60, 62, 63, 65, 67, 75, 76, 89, 91, 93, 94, 98, 104, 109, 110, 113, 114, 115, 116, 117], "return": [0, 1, 2, 3, 4, 23, 24, 29, 30, 31, 32, 33, 34, 37, 42, 43, 44, 45, 46, 54, 55, 56, 57, 58, 59, 60, 62, 64, 65, 71, 72, 75, 76, 77, 88, 89, 90, 91, 93, 94, 96, 98, 101, 102, 103, 108, 111, 112, 113, 114, 115, 117, 118, 119], "explicit": [0, 1, 2, 3, 4, 45, 46, 55, 65, 72, 75, 82, 91, 122], "delet": [0, 1, 2, 45, 46, 55], "other": [0, 1, 2, 45, 46, 52, 53, 55, 58, 62, 64, 65, 66, 70, 71, 75, 76, 81, 82, 89, 90, 94, 120], "comparis": [0, 2], "true": [0, 1, 2, 4, 46, 49, 55, 56, 60, 62, 64, 65, 70, 71, 72, 75, 76, 77, 80, 83, 89, 91, 92, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 117, 119, 120, 123], "fals": [0, 1, 2, 3, 4, 44, 45, 46, 49, 54, 62, 64, 65, 70, 71, 72, 75, 76, 77, 80, 81, 82, 83, 89, 91, 92, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 106, 107, 108, 109, 110, 111, 112, 113, 120], "struct": [1, 21, 38, 41, 45, 54, 91], "onli": [1, 3, 4, 16, 29, 44, 46, 48, 52, 54, 55, 56, 59, 60, 64, 65, 67, 68, 71, 72, 75, 76, 82, 91, 93, 94, 95, 96, 100, 102, 110, 113, 119, 120, 123], "applic": [1, 29, 46, 52, 55, 59, 64, 71, 75, 76, 89, 90, 92, 120, 123], "kcuda": [1, 46, 56, 89], "which": [1, 2, 29, 32, 37, 46, 49, 53, 54, 55, 56, 57, 58, 59, 60, 62, 63, 64, 65, 66, 71, 72, 74, 75, 76, 77, 80, 82, 83, 88, 89, 90, 91, 92, 93, 94, 95, 96, 98, 103, 104, 105, 108, 109, 110, 111, 114, 115, 116, 117, 118, 119, 120, 121, 122], "map": [1, 46, 53, 54, 55, 57, 59, 60, 65, 75, 76, 89, 91, 92, 98, 103, 114, 115, 116, 117], "kgpu": [1, 45, 46], "To": [1, 46, 52, 54, 56, 64, 66, 71, 80, 88, 89, 90, 92, 96, 102, 109, 110, 111, 114, 115, 117], "datatyp": [1, 21, 38, 45, 46, 48, 49, 50, 71, 76, 77, 90, 96, 114, 115, 117, 119], "target": [1, 33, 45, 46, 48, 49, 52, 54, 56, 58, 59, 64, 65, 66, 69, 71, 75, 76, 77, 90, 91, 92, 94, 96, 102, 122, 123], "gpu": [1, 32, 35, 37, 45, 46, 52, 64, 65, 71, 75, 76, 77, 89, 91, 92, 93, 96, 101, 104, 109, 110, 113, 114, 115, 117, 120, 122, 123], "run": [1, 37, 46, 49, 52, 53, 54, 55, 56, 57, 58, 59, 60, 64, 65, 66, 67, 68, 71, 72, 75, 76, 77, 82, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123], "kdla": [1, 45, 46, 123], "dla": [1, 45, 46, 49, 52, 64, 69, 71, 76, 77], "intern": [1, 16, 46, 60, 63, 73, 75, 82, 89], "note": [1, 46, 48, 54, 60, 62, 65, 66, 67, 75, 76, 80, 82, 89, 96, 102, 108, 114, 115, 117, 118, 123], "The": [1, 46, 48, 49, 52, 53, 54, 55, 56, 57, 58, 59, 60, 62, 63, 64, 65, 66, 71, 75, 76, 77, 80, 83, 88, 90, 91, 92, 93, 94, 95, 96, 97, 98, 100, 101, 102, 104, 105, 108, 109, 111, 113, 114, 115, 116, 117, 118, 121, 122], "valid": [1, 46, 56, 60, 62, 71, 75, 76, 94], "kcpu": [1, 46], "comparison": [1, 46], "an": [2, 3, 4, 48, 49, 52, 53, 54, 55, 56, 57, 58, 59, 60, 62, 64, 65, 66, 68, 71, 72, 74, 75, 76, 77, 80, 82, 83, 88, 89, 90, 91, 93, 94, 96, 98, 102, 103, 104, 108, 109, 110, 113, 114, 115, 116, 117, 118, 120, 121, 122], "memeori": 2, "layout": [2, 48, 70, 71, 76, 77], "store": [2, 4, 49, 52, 53, 58, 60, 64, 65, 71, 75, 76, 77, 88, 89, 93, 96, 98, 102, 111], "tensor": [2, 33, 44, 45, 48, 49, 52, 53, 54, 55, 56, 58, 60, 62, 63, 64, 65, 70, 71, 72, 75, 76, 77, 88, 89, 90, 91, 93, 94, 96, 101, 103, 108, 111, 113, 116], "kcontigu": [2, 45, 48], "contigu": [2, 48, 49, 52, 71, 76, 77], "nchw": [2, 71, 76, 77], "linear": [2, 56, 70, 76, 88, 96, 112, 119], "kchannelslast": [2, 45], "channel": [2, 76, 81], "last": [2, 55, 65, 76, 112], "nhwc": [2, 52], "memoryformat": [2, 45], "ptq": [3, 4, 15, 18, 19, 38, 50, 51, 52, 69, 71, 76, 77], "privat": [3, 4, 44, 45, 91], "algorithm": [3, 4, 29, 30, 44, 65, 74, 91, 110], "typenam": [3, 4, 29, 30, 44], "gener": [3, 4, 29, 52, 55, 58, 59, 60, 62, 64, 65, 66, 71, 72, 80, 82, 83, 86, 88, 89, 91, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 111, 112, 113, 114, 120], "int8calibr": [3, 20, 30, 40, 44, 50], "implement": [3, 4, 55, 56, 58, 63, 65, 75, 81, 89, 91, 93, 96, 98, 111, 120], "specifi": [3, 4, 33, 52, 54, 60, 64, 65, 66, 71, 76, 77, 80, 82, 90, 92, 113, 114, 115, 117, 118, 119, 121, 122], "calibr": [3, 4, 29, 30, 44, 49, 52, 71, 74, 76, 77, 89, 91], "read": [3, 4, 29, 30, 44, 80, 82, 91, 111], "nvinfer1": [3, 4, 29, 30, 44, 45, 49, 60, 91], "iint8calibr": [3, 4, 29, 30, 44, 45, 49, 71, 76, 77, 91], "iint8entropycalibrator2": [3, 4, 29, 30, 44, 91], "std": [3, 4, 22, 26, 28, 29, 30, 31, 33, 34, 37, 42, 44, 45, 47, 48, 49, 56, 89, 91, 114, 115, 117, 123], "string": [3, 4, 18, 20, 21, 22, 26, 28, 29, 30, 31, 33, 34, 37, 42, 44, 45, 49, 54, 56, 58, 60, 64, 71, 76, 80, 89, 91], "cache_file_path": [3, 4, 29, 30, 44], "8": [3, 52, 55, 63, 64, 66, 75, 76, 82, 83, 86, 89, 95, 96, 105, 108, 114, 115, 117, 118], "cach": [3, 4, 29, 30, 44, 52, 64, 65, 69, 71, 72, 74, 76, 89, 91, 99, 101, 114, 120], "getbatchs": [3, 4, 44], "noexcept": [3, 4, 44, 91], "overrid": [3, 4, 29, 30, 44, 54, 65, 91], "batch": [3, 4, 44, 64, 65, 72, 75, 91, 98, 105, 107, 112, 113, 114, 115, 117, 118, 123], "size": [3, 4, 44, 48, 49, 52, 55, 56, 64, 65, 70, 71, 72, 76, 77, 80, 89, 91, 93, 96, 98, 105, 107, 111, 112, 116, 118], "next": [3, 4, 53, 54, 58, 63, 72, 76, 80, 82, 83, 91, 94, 101, 103, 108, 112, 114, 115, 117], "alwai": [3, 4, 27, 52, 76, 82, 102, 113], "1": [3, 4, 33, 44, 45, 48, 49, 52, 54, 55, 56, 58, 60, 62, 63, 64, 65, 66, 70, 71, 72, 74, 75, 76, 77, 79, 80, 82, 83, 86, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 100, 101, 102, 104, 105, 107, 108, 109, 110, 111, 112, 113, 116, 118, 119, 121, 123], "due": [3, 4, 66, 81, 82, 104, 112], "issu": [3, 4, 64, 71, 76, 89, 103, 104, 107], "getbatch": [3, 4, 44], "void": [3, 4, 25, 26, 27, 28, 35, 36, 42, 44, 45], "bind": [3, 4, 33, 44, 75, 77, 82], "char": [3, 4, 44, 52, 89], "name": [3, 4, 31, 33, 37, 44, 54, 56, 58, 60, 65, 66, 67, 72, 74, 75, 76, 77, 82, 83, 88, 89, 92, 93, 94, 96, 102, 108, 112, 114, 115, 117, 119], "nbbind": [3, 4, 44], "Not": 3, "arrai": [3, 4, 33, 53, 54, 76, 77, 94, 96, 101, 111, 113], "pointer": [3, 4, 91], "fed": [3, 4, 48], "buffer": [3, 4, 65, 69, 96, 99, 114], "each": [3, 4, 49, 53, 55, 56, 58, 60, 64, 65, 66, 71, 72, 75, 80, 82, 89, 93, 94, 102, 108, 110, 120], "input": [3, 4, 21, 29, 33, 38, 44, 45, 47, 49, 50, 52, 53, 54, 55, 56, 58, 60, 62, 63, 64, 65, 68, 70, 71, 72, 73, 75, 76, 77, 83, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 108, 109, 110, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123], "number": [3, 4, 49, 52, 54, 55, 56, 60, 63, 64, 65, 71, 72, 76, 77, 80, 89, 90, 96, 102, 104, 105, 107, 113, 116, 122], "readcalibrationcach": [3, 4, 44], "size_t": [3, 4, 44, 91], "length": [3, 4, 44, 65, 70, 83, 104, 113], "how": [3, 4, 66, 67, 82, 84, 86, 88, 92, 93, 94, 96, 98, 100, 103, 112, 113, 114, 115, 116, 117, 118, 120], "enabl": [3, 4, 24, 49, 52, 54, 56, 57, 59, 64, 65, 66, 71, 72, 74, 75, 76, 77, 80, 98, 100, 102, 105, 107, 108, 109, 110, 111, 113, 119, 120], "use_cach": [3, 4, 30, 44, 74, 91, 104, 109, 110, 113], "set": [3, 4, 16, 21, 25, 27, 29, 32, 35, 37, 45, 46, 48, 49, 52, 53, 54, 55, 56, 57, 58, 59, 65, 66, 71, 72, 75, 76, 77, 80, 84, 87, 88, 89, 90, 91, 93, 94, 96, 101, 102, 109, 112, 113, 116, 118, 119, 120, 122, 123], "writecalibrationcach": [3, 4, 44], "write": [3, 4, 29, 30, 44, 65, 69, 82, 89, 91, 114, 115, 117], "provid": [3, 4, 49, 52, 54, 56, 58, 60, 62, 64, 65, 66, 68, 71, 72, 75, 76, 77, 82, 89, 90, 91, 92, 93, 94, 98, 99, 102, 103, 104, 108, 111, 113, 114, 115, 117, 118, 120, 121, 122], "cast": [3, 4, 55, 64, 71, 109, 110, 111, 119], "convienc": [3, 4, 49], "convert": [3, 4, 31, 32, 37, 52, 55, 56, 57, 59, 63, 64, 69, 71, 76, 77, 90, 92, 96, 99, 104, 105, 107, 111, 113, 114, 116, 120], "easili": [3, 4, 100], "assign": [3, 4, 81], "ptq_calibr": [3, 4, 45, 49, 91], "field": [3, 4, 63, 72, 76, 91], "compilespec": [3, 4, 21, 32, 37, 41, 45, 50, 56, 77, 89, 91, 123], "dataloaderuniqueptr": [4, 44], "libtorch": [4, 36, 60, 66, 68, 89, 91, 122], "dataload": [4, 29, 30, 44, 49, 74, 91, 112], "unique_ptr": [4, 30], "unqiue_ptr": 4, "A": [4, 29, 30, 32, 33, 47, 48, 54, 55, 56, 60, 65, 66, 71, 72, 76, 77, 83, 91, 93, 106, 114, 115, 117], "uniqu": [4, 90], "what": [4, 54, 55, 65, 68, 76, 82, 88, 89, 90, 104, 109, 110, 122], "make_data_load": [4, 91], "factori": [4, 29, 30, 64, 71, 91], "path": [4, 13, 14, 15, 29, 30, 52, 64, 65, 66, 67, 71, 74, 76, 88, 89, 91, 95, 98, 108, 112, 122], "find": [4, 65, 66, 67, 89, 96, 113], "whether": [4, 52, 54, 64, 65, 71, 72, 76, 81, 91, 105, 107, 120], "exist": [4, 31, 32, 37, 54, 63, 64, 65, 67, 71, 74, 76, 77, 91, 98, 116], "There": [4, 53, 54, 59, 60, 62, 63, 65, 66, 83, 88, 91, 102, 114, 115, 116, 117, 118, 120], "consum": [4, 53, 88], "macro": [5, 6, 7, 8, 9, 10, 11, 12, 15, 18, 20, 21, 42, 44, 45, 50, 51], "x": [5, 10, 33, 43, 55, 56, 66, 67, 68, 75, 77, 83, 88, 89, 93, 94, 96, 98, 103, 108, 112, 113, 114, 115, 117, 118, 119, 121], "includ": [13, 15, 16, 34, 36, 42, 43, 44, 45, 51, 52, 54, 56, 57, 58, 59, 62, 64, 65, 66, 67, 68, 71, 72, 75, 76, 80, 82, 88, 89, 91, 96, 111, 120], "parent": [14, 15, 18, 19, 20, 21], "cpp": [14, 15, 42, 43, 44, 45, 51, 55, 59, 66, 89, 91], "log": [15, 16, 19, 20, 38, 44, 50, 51, 55, 60, 64, 65, 69, 70, 71, 72, 76, 93, 94, 105, 107, 119], "emum": [16, 17], "messag": [16, 25, 26, 52, 73], "sever": [16, 26, 73, 108], "kinternal_error": [16, 42], "print": [16, 31, 44, 62, 64, 67, 71, 77, 82, 89, 92, 93, 94, 95, 96, 97, 98, 100, 101, 102, 104, 105, 107, 109, 110, 112, 113, 114, 115, 117], "error": [16, 49, 52, 53, 55, 59, 64, 65, 71, 73, 76, 77, 82, 89, 93, 118], "kerror": [16, 42], "all": [16, 42, 43, 44, 45, 49, 52, 54, 55, 56, 58, 62, 64, 65, 66, 67, 71, 73, 75, 76, 78, 82, 83, 88, 89, 90, 91, 94, 96, 109, 110, 114, 115, 116, 117, 119, 120, 122], "kwarn": [16, 42], "warn": [16, 44, 52, 60, 73, 75], "kinfo": [16, 42, 44], "info": [16, 32, 37, 45, 52, 60, 73, 75, 76, 119], "kdebug": [16, 42, 44], "debug": [16, 27, 45, 49, 52, 60, 62, 64, 71, 73, 75, 76, 77, 92, 93, 95, 96, 97, 98, 100, 102, 103, 105, 107, 112, 119], "kgraph": [16, 42, 55], "everyth": [16, 64, 71, 76], "intermedi": [16, 49, 52, 54, 64, 71, 73, 76, 77, 88, 119, 122], "graph": [16, 31, 32, 37, 45, 49, 52, 53, 54, 56, 57, 59, 60, 62, 63, 64, 65, 71, 72, 73, 76, 77, 88, 89, 93, 94, 96, 98, 100, 101, 102, 104, 105, 107, 111, 116, 118, 120], "lower": [16, 54, 63, 65, 69, 71, 72, 73, 76, 83, 93, 96, 98, 105, 107, 113, 116, 122], "phase": [16, 60, 63, 89, 94, 101, 102, 108, 118, 122], "class": [17, 29, 30, 44, 45, 46, 51, 58, 60, 64, 65, 73, 77, 82, 83, 88, 89, 90, 91, 93, 94, 96, 98, 103, 104, 108, 111, 112, 116, 118, 119], "int8_t": [17, 45], "select": [17, 29, 30, 37, 49, 52, 58, 64, 65, 66, 70, 71, 76, 77, 81, 84, 90, 91, 96, 111, 122], "capabl": [17, 45, 49, 52, 58, 71, 76, 77, 92, 94, 95], "kstandard": [17, 45, 49], "ksafeti": [17, 45], "kdla_standalon": [17, 45], "directori": [18, 19, 20, 21, 42, 43, 44, 45, 50, 66, 67, 71, 91, 98, 111, 114, 115, 117], "program": [18, 19, 20, 21, 29, 51, 52, 57, 58, 59, 69, 71, 88, 93, 98, 99, 109, 110, 114, 118], "list": [18, 19, 20, 21, 31, 49, 51, 53, 56, 58, 60, 62, 63, 65, 68, 70, 71, 72, 75, 76, 77, 86, 89, 90, 94, 96, 114, 115, 117], "level": [18, 23, 25, 26, 39, 42, 44, 50, 54, 55, 56, 59, 64, 65, 71, 76, 77, 86, 88, 94, 96, 114, 115, 117, 122], "get_is_colored_output_on": [18, 39, 42, 50], "get_logging_prefix": [18, 39, 42, 50], "get_reportable_log_level": [18, 39, 42, 50], "set_is_colored_output_on": [18, 39, 42, 50], "set_logging_prefix": [18, 39, 42, 50], "set_reportable_log_level": [18, 39, 42, 50], "torchscript": [19, 21, 38, 43, 45, 49, 50, 52, 56, 57, 58, 59, 63, 68, 71, 72, 74, 75, 76, 77, 90, 101, 114, 115, 116, 117, 118, 123], "str": [19, 43, 44, 50, 54, 64, 65, 70, 71, 74, 75, 76, 77, 94, 95, 96, 98, 112], "torch_tensorrt_major_vers": [19, 43, 50], "torch_tensorrt_minor_vers": [19, 43, 50], "torch_tensorrt_patch_vers": [19, 43, 50], "torch_tensorrt_vers": [19, 43, 50], "torchtrt_hidden": [19, 43, 50], "xstr": [19, 43, 50], "nvinfer": [20, 44], "fstream": [20, 44], "iostream": [20, 21, 44, 45, 89], "iter": [20, 44, 49, 52, 53, 64, 71, 74, 76, 77, 97, 98, 111, 112, 113], "memori": [20, 21, 44, 45, 55, 60, 71, 76, 77, 89, 90, 93, 96, 98, 101, 108, 109, 110, 113], "sstream": [20, 44], "vector": [20, 21, 33, 44, 45, 47, 48, 49, 56, 58, 76, 89, 91, 123], "templat": [20, 40, 44, 45, 50, 80, 89], "int8cachecalibr": [20, 29, 40, 44, 50], "make_int8_cache_calibr": [20, 40, 44, 50, 91], "make_int8_calibr": [20, 29, 40, 44, 50, 91], "cuda_runtim": [21, 45], "custom_class": [21, 45], "devic": [21, 33, 35, 38, 45, 49, 50, 52, 58, 64, 70, 71, 72, 74, 75, 76, 77, 90, 91, 92, 93, 96, 100, 104, 106, 109, 110, 111, 113, 116, 123], "graphinput": [21, 38, 45, 49, 50], "devicetyp": [21, 38, 45, 46, 50, 75, 76, 77, 91, 92, 96, 123], "tensorformat": [21, 38, 45, 48, 50, 76, 96], "enginecap": [21, 38, 45, 49, 50, 64, 71, 75, 76, 77, 92, 96], "dump_build_info": [21, 38, 45, 50], "get_build_info": [21, 38, 45, 50], "set_devic": [21, 38, 45, 50, 120], "check_method_operator_support": [21, 41, 45, 50], "compil": [21, 31, 37, 41, 45, 49, 50, 52, 54, 55, 56, 58, 60, 62, 65, 71, 72, 73, 75, 76, 77, 78, 80, 88, 90, 91, 92, 93, 94, 96, 97, 99, 100, 112, 114, 115, 117, 120, 123], "convert_method_to_trt_engin": [21, 41, 45, 50, 76, 77, 89, 92], "embed_engine_in_new_modul": [21, 41, 45, 50, 77], "current": [23, 54, 56, 58, 60, 62, 63, 64, 65, 66, 67, 71, 72, 76, 77, 80, 94, 96, 100, 104, 109, 110, 111, 112, 113, 120], "report": [23, 44, 75], "Is": [24, 76], "color": [24, 27, 82, 111], "output": [24, 27, 33, 49, 52, 53, 54, 55, 56, 58, 60, 62, 63, 64, 65, 66, 69, 71, 73, 75, 76, 77, 80, 82, 83, 89, 93, 94, 96, 98, 99, 100, 102, 106, 113, 114, 115, 116, 117, 118, 119, 121], "lvl": [25, 26, 42], "inform": [25, 33, 34, 36, 48, 52, 53, 56, 58, 62, 64, 65, 66, 71, 72, 73, 76, 82, 88, 89, 91, 92, 96, 98, 113, 118], "ad": [25, 52, 53, 54, 56, 62, 65, 66, 93, 96, 100], "abov": [25, 54, 56, 62, 65, 66, 73, 81, 82, 89, 96, 105, 107, 111, 119, 121], "msg": [26, 42], "add": [26, 53, 54, 55, 56, 60, 63, 66, 70, 80, 82, 87, 89, 90, 93, 94, 96], "global": [26, 52, 64, 71, 76, 89], "colored_output_on": [27, 42], "prefix": [27, 28, 42, 82], "help": [27, 52, 53, 60, 64, 65, 89, 95, 98, 108, 112, 113, 116, 120], "when": [27, 44, 45, 46, 52, 53, 55, 56, 57, 58, 59, 60, 64, 65, 66, 71, 75, 76, 77, 80, 82, 84, 88, 89, 91, 94, 96, 98, 100, 101, 102, 108, 113, 116, 118, 120], "termin": [27, 52, 89], "If": [27, 33, 53, 54, 55, 56, 62, 63, 64, 65, 66, 68, 71, 72, 76, 80, 82, 89, 90, 91, 94, 96, 98, 101, 102, 103, 108, 113, 114, 115, 117, 118, 119, 120, 122, 123], "build": [29, 30, 34, 49, 52, 53, 57, 59, 60, 63, 64, 65, 71, 75, 76, 81, 86, 89, 91, 93, 94, 96, 105, 107, 113, 118], "post": [29, 30, 49, 52, 63, 69, 89, 98], "train": [29, 30, 49, 52, 69, 70, 89, 90, 98, 113], "quantiz": [29, 30, 52, 64, 69, 74, 76, 89, 99, 114], "creat": [29, 30, 33, 52, 53, 54, 56, 58, 60, 65, 69, 76, 77, 82, 89, 93, 94, 96, 102, 111, 113, 114, 115, 117], "previous": [29, 33, 89, 93, 98, 102], "therefor": [29, 58, 65, 66, 75, 82, 89, 116, 120], "have": [29, 33, 44, 52, 53, 54, 55, 56, 60, 62, 63, 64, 65, 66, 67, 71, 72, 74, 75, 76, 77, 82, 88, 89, 90, 91, 93, 96, 99, 104, 105, 107, 111, 112, 114, 115, 116, 117, 118], "requir": [29, 49, 52, 53, 54, 55, 63, 64, 65, 66, 67, 71, 76, 77, 80, 89, 91, 94, 95, 96, 99, 101, 104, 108, 111, 112, 113, 114, 115, 117, 118, 120], "dataset": [29, 74, 91, 116], "save": [29, 44, 52, 58, 64, 65, 68, 69, 71, 75, 76, 77, 89, 90, 95, 97, 98, 101, 102, 106, 111, 113, 114, 115, 116, 117, 120, 122], "later": [29, 71, 89, 93, 102, 121, 122], "differ": [29, 55, 56, 59, 64, 65, 66, 71, 76, 80, 88, 94, 96, 98, 100, 109, 113, 116, 120, 122], "scratch": [29, 98, 102], "depend": [29, 34, 53, 59, 64, 65, 67, 68, 71, 89, 90, 104, 111, 113, 115, 117, 120], "howev": [29, 66, 80, 81, 89, 93, 94, 96, 98, 114, 115, 117, 118, 122], "network": [29, 30, 54, 60, 65, 76, 89, 91, 94, 96, 113, 114, 115, 116, 117, 123], "also": [29, 53, 54, 60, 62, 64, 66, 68, 80, 82, 83, 89, 90, 91, 93, 98, 108, 111, 112, 116], "recalibr": 29, "its": [29, 53, 56, 58, 60, 66, 75, 76, 82, 93, 96, 112, 114, 115, 117, 120, 122], "structur": [29, 46, 49, 56, 59, 60, 64, 71, 76, 80, 82, 86, 88, 96, 114, 115, 117], "chang": [29, 55, 56, 59, 62, 64, 65, 75, 76, 77, 80, 91, 93, 94, 98, 100, 101, 102, 111, 114, 115, 117, 120, 122], "respons": [29, 54, 58, 82, 120], "ensur": [29, 54, 55, 56, 62, 64, 66, 67, 71, 75, 93, 101, 108, 109, 110, 111], "By": [29, 30, 51, 56, 64, 66, 71, 80, 88, 98, 118], "entropi": [29, 30, 91], "v2": [29, 30, 82], "perform": [29, 30, 54, 62, 63, 71, 75, 76, 91, 93, 96, 108, 111, 113, 114, 115, 116, 117, 119, 120, 121, 122], "recommend": [29, 30, 65, 66, 76, 82, 89, 96, 114, 115, 117, 118], "feed": [29, 30, 89], "forward": [29, 30, 32, 33, 56, 58, 60, 64, 68, 71, 75, 76, 77, 88, 89, 90, 91, 92, 93, 94, 96, 103, 104, 108, 111, 112, 118, 119], "minmax": [29, 30, 91], "recomend": [29, 30], "nlp": [29, 30, 91], "task": [29, 30, 65, 91, 101, 116], "call": [29, 30, 32, 49, 54, 55, 58, 60, 65, 71, 72, 75, 76, 77, 82, 88, 89, 92, 94, 96, 98, 100, 103, 107, 116, 118, 120, 122], "e": [29, 30, 52, 55, 60, 65, 66, 67, 68, 72, 76, 88, 89, 91, 96, 98, 102, 114, 115, 117], "g": [29, 30, 52, 55, 65, 66, 67, 72, 76, 82, 91, 96, 102, 114, 115, 117], "iint8minmaxcalibr": [29, 30, 91], "calibration_cache_fil": [29, 30, 91], "move": [30, 44, 55, 58, 77, 89, 91, 94, 101, 109, 110], "calibration_dataload": [30, 91], "contain": [30, 31, 52, 53, 54, 55, 56, 60, 65, 66, 72, 75, 76, 82, 83, 88, 89, 91, 96, 98, 101, 111, 114, 115, 117, 120], "jit": [31, 32, 33, 37, 45, 47, 49, 52, 53, 55, 56, 57, 58, 59, 60, 61, 64, 68, 69, 71, 75, 76, 77, 88, 89, 90, 92, 93, 96, 102, 114, 115, 117, 121, 122], "modul": [31, 32, 33, 37, 45, 49, 52, 56, 57, 58, 59, 60, 64, 65, 66, 67, 68, 69, 71, 72, 74, 75, 76, 77, 81, 82, 83, 90, 91, 92, 93, 94, 95, 96, 99, 101, 102, 103, 104, 111, 112, 114, 116, 118, 119, 121, 123], "method_nam": [31, 37, 45, 52, 76, 77, 89], "see": [31, 55, 56, 58, 62, 64, 65, 66, 76, 77, 82, 88, 89, 90, 93, 94, 96, 98, 102, 103], "fulli": [31, 52, 55, 64, 71, 75, 76, 77, 89, 91, 96, 123], "take": [31, 32, 33, 37, 53, 54, 57, 58, 59, 60, 62, 65, 71, 72, 75, 76, 77, 80, 82, 89, 91, 92, 94, 96, 103, 116, 118], "method": [31, 32, 33, 37, 48, 52, 55, 60, 66, 71, 76, 77, 82, 88, 89, 92, 98, 116], "pure": [31, 71, 76], "Will": 31, "out": [31, 44, 53, 55, 56, 57, 59, 60, 64, 66, 71, 76, 77, 82, 89, 96, 100, 111, 112, 113, 114, 115, 117, 118], "unsupport": [31, 49, 54, 64, 76, 96, 101, 122], "script": [31, 55, 56, 68, 76, 77, 88, 89, 90, 92, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 117, 120, 122], "nvidia": [32, 37, 42, 43, 44, 45, 52, 61, 64, 65, 66, 67, 71, 76, 77, 89, 103, 107, 114, 115, 117, 122, 123], "configur": [32, 37, 48, 62, 64, 66, 71, 75, 76, 77, 86, 89, 91, 96, 113, 114, 115, 117, 118], "equival": [32, 57, 59, 60, 71, 76, 77, 88, 89, 91, 94, 96, 105, 107], "specif": [32, 49, 54, 55, 57, 59, 62, 64, 71, 76, 77, 82, 94, 113, 122], "traget": 32, "input_binding_nam": [33, 45, 75, 77], "output_binding_nam": [33, 45, 75, 77], "emb": [33, 52, 63, 77, 83], "pre": [33, 55, 69, 74, 77, 91, 98, 99, 113, 114, 120], "built": [33, 52, 58, 59, 64, 66, 71, 75, 76, 77, 98, 102, 111], "serial": [33, 37, 52, 57, 59, 66, 71, 75, 76, 77, 89, 96, 98, 114, 115, 117, 122], "regist": [33, 54, 58, 60, 65, 75, 77, 93, 94, 96], "execut": [33, 49, 52, 55, 57, 58, 59, 63, 64, 65, 66, 69, 71, 72, 75, 76, 77, 78, 88, 89, 91, 94, 96, 101, 108, 114, 115, 117], "must": [33, 48, 49, 52, 54, 55, 56, 60, 62, 65, 66, 71, 72, 76, 77, 82, 83, 89, 93, 98, 118, 120, 122], "follow": [33, 52, 54, 56, 58, 62, 63, 64, 65, 66, 77, 80, 82, 83, 87, 88, 89, 91, 93, 94, 96, 98, 99, 104, 105, 109, 110, 114, 115, 116, 117, 118, 119, 120], "format": [33, 45, 48, 49, 52, 70, 71, 76, 77, 82, 83, 90, 96, 98, 112, 114, 115, 116, 117, 119, 121], "symbol": [33, 65, 66, 77, 82, 120], "index": [33, 61, 62, 66, 67, 69, 70, 77, 80, 86, 91, 96, 111], "0": [33, 43, 44, 45, 49, 52, 54, 56, 59, 60, 62, 64, 65, 66, 67, 69, 70, 71, 72, 74, 75, 76, 77, 78, 79, 81, 82, 89, 91, 92, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 117, 118, 119, 123], "2": [33, 43, 54, 56, 60, 63, 64, 65, 66, 67, 69, 70, 71, 74, 75, 76, 77, 80, 82, 83, 86, 88, 89, 91, 93, 94, 96, 97, 98, 100, 102, 103, 104, 105, 107, 108, 109, 110, 111, 112, 113, 118, 121], "y": [33, 56, 77, 83, 93, 94, 96, 103], "compilesepc": 33, "order": [33, 49, 54, 56, 60, 62, 65, 66, 71, 72, 75, 76, 77, 89, 90, 94, 98, 119], "pass": [33, 53, 54, 56, 57, 58, 59, 60, 63, 64, 65, 66, 69, 73, 74, 75, 76, 77, 88, 89, 91, 93, 94, 96, 98, 102], "origin": [33, 65, 72, 76, 96, 98, 100, 111, 122], "pytorch": [33, 48, 49, 52, 54, 55, 56, 57, 58, 59, 60, 63, 64, 66, 67, 68, 71, 74, 75, 76, 77, 88, 89, 90, 91, 94, 98, 100, 101, 102, 111, 112, 114, 115, 117, 118, 119, 120, 121, 122], "assum": [33, 75, 92, 96, 99, 114], "convent": 33, "below": [33, 56, 60, 62, 63, 64, 65, 66, 67, 82, 89, 90, 98, 101, 106, 111, 114, 115, 117], "librari": [34, 42, 43, 44, 45, 52, 54, 57, 58, 59, 60, 76, 89, 93, 96, 99, 114], "version": [34, 36, 59, 62, 64, 65, 67, 71, 76, 80, 83, 96, 114, 115, 116, 117, 121], "gpu_id": [35, 45, 46, 52, 75, 76, 77, 91, 92, 96, 123], "id": [35, 45, 52, 76, 80, 81, 85, 93, 123], "cudasetdevic": 35, "dump": [36, 52, 96], "base": [36, 50, 58, 63, 64, 66, 71, 72, 76, 82, 88, 90, 91, 93, 97, 101, 102, 107, 111, 116, 122], "stdout": [36, 75], "equivil": 37, "document": [42, 43, 44, 45, 50, 59, 80, 82, 83, 87, 88, 89, 91, 92, 114, 115, 117, 118, 120], "copyright": [42, 43, 44, 45, 83, 89], "c": [42, 43, 44, 45, 52, 59, 64, 67, 70, 71, 72, 75, 76, 83, 90, 96, 100, 114, 115, 117, 120, 123], "corpor": [42, 43, 44, 45], "right": [42, 43, 44, 45, 55, 59, 60, 82, 114, 115, 117], "reserv": [42, 43, 44, 45, 109, 110], "licens": [42, 43, 44, 45, 89], "under": [42, 43, 44, 45, 59, 65, 82, 94, 105, 122], "bsd": [42, 43, 44, 45], "style": [42, 43, 44, 45, 64, 68, 80, 82, 83], "found": [42, 43, 44, 45, 63, 66, 75, 82, 89, 91, 94, 96, 98, 120], "root": [42, 43, 44, 45, 66, 80, 91, 112], "sourc": [42, 43, 44, 45, 54, 59, 64, 65, 67, 71, 72, 73, 74, 75, 76, 77, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114], "tree": [42, 43, 44, 45, 80, 91, 112, 120], "pragma": [42, 43, 44, 45, 91], "onc": [42, 43, 44, 45, 53, 55, 56, 58, 64, 65, 66, 67, 76, 91, 96, 110, 113, 114, 115, 117, 120], "namespac": [42, 43, 44, 45, 51, 55, 69, 76, 91, 93, 96], "ar": [42, 46, 49, 52, 53, 54, 55, 56, 58, 59, 60, 62, 63, 64, 65, 66, 71, 74, 75, 76, 77, 80, 82, 83, 84, 88, 89, 91, 92, 93, 94, 96, 97, 98, 100, 101, 102, 105, 108, 109, 110, 111, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122], "ones": [42, 56, 57, 59, 66, 82, 89, 94, 96, 122], "necessari": [42, 62, 64, 66, 75, 93, 94, 102, 120], "user": [42, 48, 54, 56, 57, 58, 59, 62, 63, 64, 66, 67, 71, 82, 83, 89, 90, 91, 94, 98, 102, 113, 114, 115, 117, 118, 119, 120, 122], "dont": 42, "know": [42, 60, 80, 82, 93, 94, 96, 104], "we": [42, 44, 53, 54, 55, 56, 57, 58, 59, 60, 62, 63, 64, 65, 72, 75, 80, 82, 88, 89, 91, 93, 94, 96, 98, 99, 100, 101, 102, 103, 104, 105, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 121, 122], "want": [42, 56, 65, 66, 67, 68, 72, 88, 89, 91, 92, 94, 96, 102, 103, 114, 115, 117], "use_cmake_generated_export_head": 43, "torch_tensorrt_export": 43, "els": [43, 44, 48, 77, 82, 83, 95, 97, 98, 111, 112], "__gnuc__": 43, "__attribute__": 43, "__visibility__": 43, "hidden": [43, 80], "endif": [43, 44, 45], "doe": [43, 44, 55, 56, 60, 62, 65, 66, 76, 82, 91, 93, 96, 105, 107], "gaurd": 43, "someth": [43, 55, 82, 114, 115, 117], "6": [43, 55, 56, 58, 66, 70, 82, 86, 88, 89, 95, 96, 111], "setup": [43, 67, 91, 114, 115, 117], "alias": 43, "eas": 43, "ts": [43, 52, 56, 68, 69, 76, 88, 89, 90, 92, 118, 121], "torchtrt": [43, 56, 95, 96, 112, 114, 115, 117], "ifndef": [44, 45], "doxygen_should_skip_thi": [44, 45], "get_batch_impl": 44, "element_typ": 44, "super": [44, 88, 93, 94, 96, 103, 111, 112, 118, 119], "batchtyp": 44, "dataloader_": 44, "cache_file_path_": 44, "use_cache_": 44, "auto": [44, 56, 60, 64, 68, 71, 82, 83, 89, 91, 104, 109, 110, 113, 123], "batched_data_": 44, "push_back": [44, 56], "it_": 44, "begin": [44, 65, 66, 82, 103, 108], "hack": 44, "explict": 44, "work": [44, 55, 59, 60, 64, 65, 68, 71, 74, 75, 76, 82, 83, 91, 93, 96, 102, 103, 108, 113, 114, 115, 117, 118], "here": [44, 53, 54, 56, 58, 63, 64, 65, 66, 68, 80, 82, 83, 88, 89, 91, 93, 94, 96, 99, 108, 109, 110, 111, 112, 114, 115, 117, 118, 120, 121], "explic": 44, "just": [44, 45, 55, 56, 64, 65, 69, 73, 75, 82, 84, 88, 89, 90, 92, 93, 96, 98, 100, 116, 120], "still": [44, 56, 65, 66, 91, 94, 103, 122], "static_cast": 44, "option": [44, 48, 52, 56, 57, 59, 62, 63, 64, 65, 71, 75, 76, 77, 82, 86, 91, 94, 96, 97, 98, 103, 104, 106, 108, 119, 120, 121, 123], "batch_siz": [44, 91, 112], "end": [44, 52, 60, 62, 70, 71, 76, 77, 82, 89, 91, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113], "statu": [44, 83], "reset": [44, 97, 98, 103, 107, 120], "incas": 44, "go": [44, 55, 56, 65, 68, 88, 89, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 122], "again": [44, 58, 60, 82, 96, 100], "stringstream": 44, "ss": 44, "cache_": 44, "clear": 44, "ifstream": 44, "io": [44, 67, 114, 115, 117], "binari": [44, 91], "noskipw": 44, "good": [44, 60, 65, 82, 98], "copi": [44, 60, 65, 67, 70, 74, 83, 113], "istream_iter": 44, "back_insert": 44, "nullptr": [44, 45, 49], "ofstream": [44, 89], "cache_fil": [44, 74, 91], "reinterpret_cast": 44, "cache_size_": 44, "arrayref": [45, 48, 49], "friend": 45, "ostream": 45, "os": [45, 67, 98], "dtype": [45, 48, 49, 52, 63, 64, 65, 70, 71, 72, 75, 76, 77, 90, 93, 96, 97, 101, 105, 107, 108, 111, 113, 114, 115, 116, 117, 118, 119], "device_typ": [45, 46, 76, 91, 92, 123], "int64_t": [45, 46, 48, 49, 91, 123], "core": [45, 52, 55, 56, 59, 64, 71, 76, 89, 94, 122, 123], "agx": 45, "platform": [45, 52, 59, 64, 66, 67, 71, 95, 123], "xavier": [45, 123], "dla_cor": [45, 46, 52, 76, 91, 92, 123], "allow_gpu_fallback": [45, 46, 71, 76, 77, 91, 92, 123], "customclasshold": [45, 48], "min_shap": [45, 48, 63, 65, 71, 76, 77, 90, 105, 108, 116, 118], "opt_shap": [45, 48, 63, 71, 76, 77, 90, 105, 108, 116, 118], "max_shap": [45, 48, 63, 65, 71, 76, 77, 90, 105, 108, 116, 118], "shape": [45, 47, 48, 49, 52, 56, 60, 63, 65, 69, 70, 71, 72, 75, 76, 77, 78, 90, 93, 94, 96, 99, 101, 108, 111, 112, 113, 114, 115, 117, 120, 123], "doubl": [45, 48, 49, 52, 63, 71, 76, 77, 82, 120], "tensor_domain": [45, 48, 76], "input_is_dynam": 45, "ivalu": [45, 47, 49, 53, 58, 60, 89], "input_signatur": [45, 47, 49, 77, 90], "nest": [45, 49, 50, 82, 83], "full": [45, 49, 52, 60, 64, 71, 73, 76, 89, 91, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 117, 120, 123], "spec": [45, 48, 49, 52, 73, 76, 77, 92, 98], "flatten": [45, 47, 70, 88, 89, 112], "fixed_s": [45, 49], "reflect": [45, 76], "builderconfig": 45, "graph_input": [45, 49], "enabled_precis": [45, 49, 63, 64, 71, 75, 76, 77, 89, 90, 91, 92, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 109, 110, 111, 112, 113, 114, 115, 117, 119, 123], "disable_tf32": [45, 49, 64, 71, 75, 76, 77, 91, 96, 104, 109, 110], "sparse_weight": [45, 49, 64, 65, 71, 75, 76, 77, 96], "refit": [45, 49, 64, 69, 71, 76, 77, 92, 96, 98, 99, 100, 114], "truncate_long_and_doubl": [45, 49, 63, 64, 77, 106], "allow_shape_tensor": [45, 49, 77], "uint64_t": [45, 49], "num_avg_timing_it": [45, 49, 64, 71, 75, 76, 77, 92, 96], "workspace_s": [45, 49, 52, 64, 71, 75, 76, 77, 96, 102, 105, 107], "dla_sram_s": [45, 49, 52, 64, 71, 75, 76, 77, 96], "1048576": [45, 49, 64, 71, 75, 76, 77, 96], "dla_local_dram_s": [45, 49, 52, 64, 71, 75, 76, 77, 96], "1073741824": [45, 49, 64, 71, 75, 76, 77, 96], "dla_global_dram_s": [45, 49, 52, 64, 71, 75, 76, 77, 96], "536870912": [45, 49, 64, 71, 75, 76, 77, 96], "require_full_compil": [45, 49, 64, 71, 75, 76, 77, 96], "min_block_s": [45, 49, 56, 63, 64, 71, 75, 76, 77, 93, 94, 95, 96, 97, 98, 102, 103, 104, 105, 107, 108, 111, 112], "3": [45, 49, 52, 55, 56, 58, 63, 64, 65, 67, 68, 70, 71, 74, 76, 77, 82, 83, 86, 88, 89, 91, 92, 93, 95, 96, 97, 98, 100, 101, 102, 105, 108, 109, 110, 111, 112, 113, 116, 118, 121, 123], "torch_executed_op": [45, 49, 56, 63, 64, 71, 75, 76, 77, 96, 102, 103, 105, 107, 108], "torch_executed_modul": [45, 49, 56, 71, 76, 77], "member": [46, 47, 48, 49], "hold": [46, 47, 48, 53, 60, 76, 91], "relat": [46, 82, 103, 107], "let": [46, 52, 55, 60, 65, 71, 76, 77, 80, 82, 114, 115, 116, 117, 122], "layer": [46, 49, 52, 53, 55, 60, 62, 64, 65, 71, 75, 76, 77, 89, 91, 94, 96, 109, 110, 112, 114, 115, 116, 117, 118, 119, 122, 123], "thei": [46, 52, 53, 54, 55, 58, 60, 64, 65, 71, 74, 75, 76, 80, 82, 90, 94, 98], "complex": [47, 49, 64, 66, 88, 90, 93, 100, 110], "either": [47, 48, 52, 60, 62, 71, 76, 77, 80, 82, 88, 89, 90, 93, 94, 95, 96, 98, 121], "one": [47, 54, 55, 60, 64, 65, 67, 71, 75, 76, 82, 88, 89, 90, 93, 94, 96, 103, 107, 109, 110, 114, 115, 117], "rang": [48, 49, 52, 65, 76, 93, 96, 97, 98, 101, 104, 105, 113, 116, 118], "optim": [48, 52, 63, 64, 65, 69, 71, 72, 74, 76, 88, 89, 90, 101, 102, 104, 105, 106, 107, 108, 111, 113, 116, 118, 122], "profil": [48, 72, 75, 119], "singl": [48, 52, 55, 56, 65, 76, 82, 88, 89, 91, 108, 111, 113, 120], "repres": [48, 49, 54, 60, 65, 68, 82, 101, 111], "signifi": [48, 55], "static": [48, 49, 53, 60, 63, 64, 71, 76, 77, 80, 89, 101, 112, 118], "three": [48, 57, 59, 65, 72, 76, 82, 83, 114, 115, 116, 117], "min": [48, 52, 60, 70, 76, 98, 104, 105, 118], "optimin": 48, "max": [48, 52, 60, 70, 76, 80, 98, 104, 105, 112, 118], "allow": [48, 49, 52, 53, 54, 55, 56, 62, 64, 65, 66, 71, 76, 77, 80, 93, 94, 96, 98, 101, 102, 105, 107, 108, 113, 120], "argument": [48, 52, 54, 55, 58, 60, 62, 64, 65, 71, 75, 76, 77, 82, 83, 89, 90, 94, 95, 96, 118], "expect": [48, 54, 55, 60, 76, 89, 90, 93, 116], "tradit": [48, 71, 76, 77, 91], "convect": 48, "produc": [48, 53, 54, 58, 60, 63, 76, 82, 89, 116], "low": [48, 65, 94, 100, 111], "high": [48, 55, 56, 80, 94, 96, 122], "weight": [48, 49, 52, 53, 64, 65, 69, 70, 71, 75, 76, 77, 82, 89, 98, 99, 100, 106, 114, 116], "first": [48, 53, 54, 55, 65, 68, 82, 83, 89, 90, 91, 94, 96, 98, 100, 103, 104, 114, 115, 117, 118, 121, 122], "calcul": [48, 53, 56, 89, 96, 113], "detect": [48, 58, 76], "float32": [48, 49, 52, 63, 64, 65, 71, 76, 77, 96, 100, 104, 106, 109, 110, 113, 118, 119], "dynam": [48, 49, 63, 65, 69, 71, 72, 76, 77, 78, 94, 98, 99, 103, 104, 106, 107, 110, 113, 114, 120], "opt": [48, 66, 75, 76, 108], "minimum": [48, 49, 52, 56, 63, 64, 71, 76, 77, 96, 113], "maximum": [48, 49, 52, 64, 65, 71, 72, 76, 77, 104, 105, 107, 113, 114, 115, 117], "accept": [48, 52, 54, 58, 60, 66, 76, 89, 90, 103, 121], "exampl": [48, 56, 58, 59, 60, 65, 66, 71, 73, 75, 76, 77, 78, 80, 81, 83, 86, 88, 89, 90, 91, 93, 94, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 120, 121], "s": [48, 49, 53, 56, 58, 60, 63, 65, 66, 67, 69, 71, 72, 75, 76, 80, 82, 83, 88, 89, 91, 94, 96, 98, 111, 113, 114, 115, 116, 117, 118, 120, 121], "cannot": [48, 55, 56, 65, 66, 71, 75, 76, 77, 81, 88, 93, 95, 96, 101], "through": [48, 53, 54, 55, 56, 58, 64, 65, 71, 73, 74, 82, 89, 90, 96, 100, 102, 116, 122], "altern": [48, 56, 62, 63, 76, 90, 94, 101, 108, 116, 121], "refer": [48, 54, 57, 59, 65, 81, 86, 89, 91, 96, 112, 114, 115, 117, 118, 121], "given": [48, 49, 52, 54, 55, 65, 71, 72, 74, 76, 77, 88, 89, 90, 92, 93, 94, 111, 118], "kernel": [48, 49, 52, 60, 64, 65, 69, 71, 76, 77, 94, 99, 108, 114, 119, 120], "ani": [48, 52, 53, 54, 60, 62, 64, 65, 70, 71, 74, 75, 76, 77, 80, 82, 89, 90, 91, 94, 96, 105, 118], "event": [48, 64, 97, 98], "place": [48, 55, 62, 65, 82, 83, 84, 91, 93, 96, 112], "variabl": [48, 65, 75, 76], "dimens": [48, 55, 65, 72, 76, 105, 116, 118, 119], "domain": [48, 76, 83, 91], "convien": 49, "fix": [49, 65, 82, 93, 96, 120, 123], "describ": [49, 56, 60, 76, 88, 92, 93, 114, 115, 117], "entri": [49, 60, 98], "okai": 49, "ha": [49, 53, 54, 55, 56, 57, 59, 60, 62, 64, 65, 66, 67, 71, 72, 75, 76, 82, 83, 88, 89, 91, 94, 95, 98, 101, 102, 108, 112, 116, 118, 122], "flaten": 49, "precis": [49, 52, 63, 64, 65, 69, 71, 76, 89, 90, 91, 105, 107, 109, 110, 111, 113, 123], "dure": [49, 52, 54, 56, 60, 63, 64, 71, 74, 76, 91, 94, 108, 109, 110, 113, 114, 115, 116, 117, 118, 120], "prevent": [49, 52, 54, 56, 108], "tf32": [49, 52, 64, 71], "comput": [49, 64, 65, 66, 67, 71, 75, 82, 91, 93, 95, 99, 101, 114, 116], "inner": [49, 83, 116], "product": [49, 67, 76], "round": [49, 71, 76, 77, 96], "10": [49, 66, 67, 71, 72, 76, 77, 86, 88, 89, 91, 93, 101, 111, 112, 113, 114, 115, 116, 117, 118, 119], "bit": [49, 60, 65, 66, 71, 76, 77, 89], "mantissa": [49, 71, 76, 77], "befor": [49, 54, 55, 56, 59, 60, 65, 71, 76, 77, 89, 101, 104, 111, 114, 115, 117, 118], "multipli": [49, 71, 76, 77], "accumul": [49, 64, 71, 76, 77, 109, 110, 111], "sum": [49, 65, 70, 71, 76, 77, 96, 112], "23": [49, 55, 71, 76, 77, 83], "behavior": [49, 56, 65, 71, 76, 77, 94, 109, 110, 118, 120, 121], "sparsiti": [49, 52, 65, 71, 76, 77], "conv": [49, 52, 89, 96], "fc": [49, 52, 55], "truncat": [49, 52, 63, 64, 71, 76, 77], "long": [49, 52, 53, 63, 76, 82, 83, 93], "float": [49, 52, 63, 64, 70, 76, 88, 89, 90, 91, 92, 93, 96, 97, 98, 102, 103, 107, 108, 111, 119], "ishap": 49, "restrict": [49, 64, 71, 76, 77, 118], "cuda": [49, 58, 63, 65, 67, 68, 71, 72, 75, 76, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 100, 101, 102, 104, 105, 106, 109, 110, 111, 112, 113, 114, 115, 117, 118, 119, 120, 121], "safeti": [49, 52, 76], "averag": [49, 52, 64, 71, 76, 77, 96], "time": [49, 52, 53, 54, 55, 56, 57, 58, 59, 60, 64, 65, 66, 68, 69, 71, 72, 75, 76, 77, 80, 82, 89, 91, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113], "workspac": [49, 52, 64, 65, 66, 71, 72, 76, 77, 96, 103, 105, 107], "fast": [49, 52, 64, 68, 71, 76, 77], "softwar": [49, 52, 64, 71, 76, 77, 82], "manag": [49, 52, 53, 55, 57, 59, 60, 64, 66, 67, 71, 73, 75, 76, 77, 89, 101, 108, 120], "ram": [49, 52, 64, 71, 76, 77], "commun": [49, 52, 64, 71, 76, 77, 89], "within": [49, 52, 57, 59, 64, 69, 71, 75, 76, 77, 80, 82, 93, 99, 108, 109, 110, 114, 116], "host": [49, 52, 64, 66, 71, 76, 77, 93, 96, 113, 114, 115, 117], "share": [49, 52, 64, 66, 71, 75, 76, 77, 98], "across": [49, 52, 55, 56, 64, 71, 76, 77, 80, 101], "metadata": [49, 52, 54, 58, 60, 64, 71, 76, 77, 80, 102, 118, 119], "quantizatiom": 49, "instead": [49, 52, 53, 54, 55, 66, 71, 75, 76, 89, 94, 102, 111, 112, 120], "potenti": [49, 71, 76, 85, 101], "subgraph": [49, 52, 53, 54, 55, 60, 62, 89, 96, 98, 101, 122], "aten": [49, 54, 55, 56, 60, 61, 64, 69, 70, 71, 76, 77, 89, 94, 103, 108, 122], "thrown": [49, 71, 76, 77], "empti": [49, 71, 72, 76, 77, 83, 88, 96, 111], "torch_tensorrtnamespac": 50, "loggingenum": 50, "levelnamespac": 50, "ptqtemplat": 50, "int8cachecalibratortempl": 50, "int8calibratornamespac": 50, "torchscriptstruct": 50, "compilespecstruct": 50, "deviceclass": 50, "devicetypestruct": 50, "graphinputsstruct": 50, "inputclass": 50, "datatypeclass": 50, "tensorformatenum": 50, "cppdirectori": 50, "includedirectori": 50, "torch_tensorrtfil": 50, "hfile": 50, "relationship": 50, "inherit": [50, 65, 71, 91], "subdirectori": 51, "definit": [51, 54, 60, 82], "cli": [52, 90], "It": [52, 54, 55, 56, 57, 59, 60, 65, 66, 69, 76, 80, 82, 93, 95, 96, 113, 116, 120, 122], "serv": [52, 58, 65, 69, 71, 76], "easi": [52, 53, 55, 89, 91], "wai": [52, 64, 65, 66, 88, 89, 91, 93, 94, 96, 98, 102, 116, 120, 121], "command": [52, 64, 66, 82, 83, 88, 89, 114, 115, 117], "line": [52, 66, 83, 89, 100], "quickli": [52, 89, 91, 114, 115, 117], "part": [52, 56, 59, 65, 75, 80, 81, 82, 93, 96, 98, 101], "deploy": [52, 75, 89, 90, 91, 93, 114, 115, 116, 117, 120, 123], "pipelin": [52, 89, 100, 106, 123], "basic": [52, 56, 65, 83, 114, 115, 117], "featur": [52, 56, 65, 66, 89, 91, 92, 106, 111, 112, 113, 116, 122], "though": [52, 59, 60, 88, 89, 122], "alreadi": [52, 53, 54, 55, 89, 91, 93, 94, 96, 99, 111, 114, 115, 117, 118], "two": [52, 55, 60, 62, 64, 65, 66, 76, 82, 83, 87, 88, 90, 91, 94, 98, 111, 114, 115, 117, 118], "embed": [52, 54, 58, 70, 77, 82, 123], "plan": [52, 59, 63, 64, 71], "after": [52, 53, 55, 56, 62, 65, 71, 75, 76, 88, 89, 90, 101, 103, 107, 114, 115, 117, 120], "link": [52, 53, 62, 69, 80, 81, 86, 89, 96, 120], "against": [52, 89, 94], "libtorchtrt": [52, 66, 89], "python": [52, 56, 59, 62, 64, 65, 67, 71, 72, 75, 76, 77, 82, 83, 89, 92, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 120, 123], "import": [52, 55, 56, 63, 64, 65, 66, 67, 68, 75, 80, 82, 88, 89, 90, 92, 93, 94, 96, 97, 98, 100, 114, 115, 117, 118, 120, 121], "packag": [52, 55, 64, 67, 89], "aspect": 52, "ident": [52, 62, 64, 71, 76, 93, 102], "standard": [52, 58, 66, 69, 71, 75, 76, 77, 82, 92, 93, 94, 96, 100, 111, 116, 120], "load": [52, 56, 58, 64, 65, 68, 71, 74, 75, 76, 77, 89, 90, 91, 92, 93, 95, 96, 97, 98, 100, 102, 104, 113, 114, 115, 116, 117, 120, 122], "like": [52, 53, 55, 58, 60, 65, 66, 68, 76, 81, 82, 88, 89, 90, 91, 93, 94, 96, 98, 100, 102, 104, 113, 114, 115, 117, 120], "would": [52, 54, 60, 64, 65, 66, 67, 75, 89, 90, 92, 94, 96, 104, 114, 115, 117, 120], "input_file_path": [52, 123], "output_file_path": [52, 123], "input_spec": [52, 65, 72], "displai": [52, 62, 64, 73, 80, 120], "menu": [52, 80, 82], "verbios": 52, "v": [52, 67, 83, 112, 114, 115, 117], "verbos": [52, 64, 65, 71, 72, 83, 105, 107], "about": [52, 53, 58, 60, 66, 75, 80, 89, 111, 114, 115, 117, 118], "process": [52, 56, 64, 76, 81, 82, 88, 91, 92, 93, 101, 102, 103, 108, 114, 115, 116, 117, 120], "onto": [52, 58], "consol": 52, "w": [52, 66, 76, 111], "disabl": [52, 64, 66, 71, 75, 80, 81, 94, 98, 113, 120], "i": [52, 55, 60, 66, 68, 70, 82, 83, 88, 89, 91, 93, 96, 97, 98, 101, 104, 109, 111, 112], "debugg": [52, 71, 76, 77], "fallback": [52, 57, 59, 60, 101, 102, 123], "model": [52, 56, 58, 63, 68, 71, 72, 73, 74, 76, 78, 88, 89, 90, 91, 92, 97, 98, 100, 118, 120, 122], "throw": [52, 55, 76, 89], "spars": [52, 54, 64, 70, 71], "p": [52, 70, 89, 114, 115, 117, 123], "repeat": [52, 70], "f32": [52, 71, 75, 76, 96], "half": [52, 64, 76, 82, 89, 90, 91, 92, 96, 101, 103, 105, 109, 110, 111, 113, 119, 123], "float16": [52, 76, 96, 100, 106, 111, 119], "f16": [52, 76, 89, 114, 115, 117, 123], "i8": [52, 76], "d": [52, 67, 76, 82, 83, 89, 123], "multi": [52, 75], "dlacor": 52, "avail": [52, 54, 60, 62, 64, 65, 66, 67, 71, 75, 76, 80, 96, 104, 111, 113, 116, 122, 123], "dla_standalon": [52, 76], "file_path": [52, 76, 95, 121], "teo": 52, "op_nam": 52, "op": [52, 53, 54, 55, 56, 57, 59, 60, 62, 63, 64, 75, 76, 89, 93, 94, 103, 108, 120, 122], "partial": [52, 82], "tem": 52, "module_nam": 52, "mod": [52, 56, 65, 71, 86, 89, 91, 119], "mb": [52, 78], "num_op": 52, "block": [52, 53, 55, 56, 64, 71, 86, 93, 122], "treat": 52, "num": 52, "avg": 52, "num_it": 52, "sram": 52, "local": [52, 55, 66, 67, 80, 89], "dram": 52, "atol": 52, "absolut": [52, 66], "toler": 52, "threshold": 52, "numer": [52, 65, 83], "deviat": 52, "1e": [52, 100, 102], "rtol": 52, "rel": [52, 56, 101], "5": [52, 56, 58, 59, 64, 65, 66, 67, 71, 75, 76, 82, 83, 86, 88, 89, 94, 96, 100, 101, 103, 108, 111, 113, 114, 115, 117], "skip": 52, "complianc": 52, "64bit": [52, 95], "32bit": 52, "custom": [52, 62, 63, 65, 66, 69, 99, 109, 110, 111, 114], "dll": 52, "n": [52, 60, 62, 76, 89, 91, 93, 94, 96, 97], "min_n": 52, "min_c": 52, "min_h": 52, "min_w": 52, "opt_n": 52, "opt_c": 52, "opt_h": 52, "opt_w": 52, "max_n": 52, "max_c": 52, "max_h": 52, "max_w": 52, "32": [52, 76, 88, 89, 90, 91, 104, 109, 110, 112, 123], "flag": [52, 56, 57, 59, 64, 66, 71, 74, 76, 90, 108, 109, 110, 120, 121], "forc": [52, 63, 65, 71, 76, 77, 80, 111], "posit": [52, 54, 65, 76, 80], "test": [52, 56, 59, 65, 66, 67, 71, 76, 82, 83, 91, 112, 114, 115, 116, 117], "ssd_trace": 52, "pt": [52, 65, 89, 104, 109, 110, 114, 115, 117], "ssd_trt": 52, "300": [52, 92, 93], "512": [52, 71, 76, 77, 112, 116], "1024": [52, 71, 76, 77, 93, 109, 116], "simplifi": [53, 96], "form": [53, 75, 76, 82, 90, 114, 115, 117], "up": [53, 55, 56, 57, 58, 59, 62, 65, 66, 71, 76, 82, 88, 93, 94, 96, 98, 101, 102, 103, 107, 108, 113, 116], "context": [53, 57, 58, 59, 64, 73, 75, 94, 101, 104, 108, 111, 120], "inetworkdefinit": [53, 54], "record": [53, 88, 97, 98, 108, 120], "togeth": [53, 60, 89], "start": [53, 56, 65, 70, 74, 76, 83, 89, 92, 96, 97, 98, 104, 116], "look": [53, 54, 55, 68, 71, 76, 88, 91, 92, 94, 98, 104, 114, 115, 117, 118], "assembl": [53, 62, 89], "resourc": [53, 91, 93, 96, 101], "coupl": [53, 59, 65, 120], "state": [53, 54, 60, 62, 75, 89, 94, 100, 104, 111], "been": [53, 60, 64, 66, 67, 74, 83, 89, 95, 98, 101, 102, 111, 122], "evaluated_value_map": [53, 60], "stage": [53, 65], "arg": [53, 54, 62, 65, 71, 74, 75, 76, 86, 89, 94, 95, 96, 98, 112, 116], "itensor": [53, 54, 60, 65, 89, 94, 96], "value_tensor_map": [53, 60], "typic": [53, 60, 76, 101, 108, 114, 115, 117], "abl": [53, 55, 60, 62, 65, 91, 92, 96, 102, 104], "system": [53, 60, 62, 64, 69, 71, 75, 76, 77, 94, 95, 96, 98, 102, 122], "registri": [53, 54, 89, 96], "enter": [53, 76], "recurs": 53, "resolv": [53, 55, 57, 59, 103, 104, 107], "until": [53, 56, 59, 60, 66, 71, 76, 122], "final": [53, 56, 57, 59, 66, 94, 96, 103, 107, 116], "some": [53, 54, 55, 56, 57, 58, 59, 60, 62, 63, 64, 65, 66, 71, 76, 81, 82, 89, 91, 93, 94, 96, 98, 101, 108, 118, 122], "These": [53, 54, 56, 58, 62, 64, 66, 71, 74, 75, 76, 80, 82, 91, 93, 94, 114, 115, 117, 122], "those": [53, 54, 62, 64, 82], "do": [53, 54, 55, 56, 60, 63, 65, 81, 83, 88, 89, 90, 91, 93, 94, 96, 109, 110, 123], "theori": [53, 82], "kind": [53, 65], "common": [53, 55, 65, 72, 82, 94, 98], "prim": [53, 55, 56, 58, 70, 88, 89], "constant": [53, 54, 55, 56, 89, 96, 101], "emit": 53, "listconstruct": [53, 56, 58, 89], "make": [53, 54, 65, 66, 67, 71, 76, 82, 84, 89, 90, 91, 96, 98, 114, 115, 116, 117, 123], "associ": [53, 60, 89, 98, 120], "where": [53, 54, 55, 60, 62, 64, 65, 71, 75, 76, 77, 83, 89, 91, 93, 94, 102, 108], "result": [53, 55, 56, 66, 68, 71, 73, 75, 76, 77, 80, 88, 90, 93, 95, 96, 100, 101, 102, 108, 113, 114, 115, 117, 119, 122], "done": [53, 56, 59, 96, 102, 114, 115, 117, 121], "mai": [53, 54, 56, 58, 59, 65, 66, 71, 75, 76, 77, 82, 83, 88, 89, 90, 91, 94, 96, 102, 103, 107, 108, 113, 114, 115, 117, 120], "For": [53, 56, 62, 63, 64, 65, 66, 68, 72, 76, 80, 82, 83, 88, 89, 91, 92, 93, 94, 96, 100, 103, 112, 114, 115, 116, 117, 120, 121], "more": [53, 64, 65, 66, 67, 69, 71, 76, 80, 83, 88, 89, 90, 91, 92, 96, 98, 100, 101, 105, 107, 111, 114, 115, 117, 120], "writing_convert": [53, 89], "locat": [54, 62, 66, 91, 94, 96], "py": [54, 55, 59, 62, 65, 66, 67, 78, 80, 82, 87, 88, 89, 91, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 117, 118], "convers": [54, 55, 56, 58, 63, 64, 65, 71, 76, 77, 89, 93, 94, 96, 111, 116, 118], "decror": 54, "dynamo_tensorrt_convert": [54, 94, 96], "signatur": [54, 77], "leaky_relu": [54, 70], "def": [54, 62, 65, 82, 88, 90, 93, 94, 96, 97, 98, 101, 103, 108, 111, 112, 113, 114, 115, 117, 118, 119], "leaky_relu_convert": 54, "ctx": [54, 60, 89, 94, 96, 113], "conversionctx": [54, 60, 89, 94], "tupl": [54, 58, 63, 65, 71, 72, 75, 76, 77, 90, 93, 94, 96, 98, 102, 118, 119], "kwarg": [54, 65, 71, 74, 75, 76, 94, 96, 116], "dict": [54, 71, 75, 76, 77, 94, 96, 98], "union": [54, 60, 64, 71, 75, 76, 77, 89, 94], "sequenc": [54, 62, 65, 71, 72, 75, 76, 77, 82, 94, 96, 104, 108, 113, 116], "decor": [54, 62, 65, 94], "kei": [54, 82, 88, 98, 114, 115, 117, 118], "node": [54, 55, 56, 57, 59, 60, 62, 64, 65, 71, 72, 89, 94, 96, 112, 116, 118], "capability_valid": [54, 94], "lambda": [54, 60, 82, 89, 93, 94, 114, 115, 117], "fx": [54, 62, 63, 71, 75, 76, 89, 90, 94, 96, 102, 121], "determin": [54, 55, 64, 65, 76, 93, 94, 113, 118, 120], "properli": [54, 66], "handl": [54, 55, 56, 58, 64, 65, 75, 76, 93, 96, 101, 108], "partition": [54, 71, 76, 96], "sure": [54, 66, 67, 89, 90, 104, 114, 115, 117, 123], "prioriti": [54, 94], "develop": [54, 65, 66, 67, 69, 82, 83, 89, 94, 96], "bodi": [54, 82, 83], "nativ": [54, 59, 61, 89, 93, 94, 96, 102], "numpi": [54, 76, 96, 97, 98, 100, 101, 102, 111, 113, 114, 115, 117], "frozen": 54, "attribut": [54, 55, 56, 58, 65, 76, 82, 89], "previou": [54, 80, 103, 111], "correspond": [54, 60, 65, 66, 75, 76, 94, 98, 100, 104, 112, 120], "edg": [54, 82], "well": [54, 63, 66, 69, 73, 75, 82, 89, 91, 93, 94, 98, 108, 111, 121], "being": [54, 65, 66, 71, 89, 94, 96, 102, 108], "truth": 54, "http": [54, 61, 64, 66, 67, 80, 82, 88, 89, 91, 94, 96, 100, 103, 107, 111, 112, 114, 115, 116, 117, 118, 120], "github": [54, 61, 64, 66, 67, 80, 89, 91, 103, 107, 111, 112, 114, 115, 117, 120], "com": [54, 61, 64, 66, 67, 89, 91, 100, 103, 107, 111, 112, 114, 115, 117, 120], "blob": [54, 61, 66, 80, 91, 98, 111], "main": [54, 55, 56, 57, 58, 59, 60, 63, 65, 66, 80, 82, 84, 89, 94, 96, 109, 111, 112], "src": [54, 58, 61, 70], "native_funct": [54, 61], "yaml": [54, 61], "sinc": [54, 55, 64, 65, 67, 75, 82, 88, 89, 91, 93, 94, 97, 98, 102, 111], "mani": [54, 56, 64, 65, 80, 82, 83, 94, 98, 102, 122], "composit": [54, 89], "raw": [54, 80, 94], "impl": [54, 93, 94], "subpackag": 54, "chain": [54, 60], "primarili": [54, 59, 66, 89, 94], "manipul": [54, 62, 76], "net": [54, 60, 82, 83, 89, 96, 114, 115, 117], "addit": [54, 55, 64, 65, 75, 76, 89, 94, 96, 98, 102, 108, 111, 116, 118], "call_modul": 54, "call_funct": [54, 62, 65], "eg": [54, 114, 115, 117, 119], "aten_": 54, "_leaky_relu": 54, "opoverloadpacket": 54, "while": [54, 56, 66, 75, 91, 94, 100, 101, 113, 114, 115, 116, 117, 120, 122], "opoverload": 54, "particular": [54, 64, 98], "collect": [54, 56, 64, 71, 76, 77, 89, 90, 112], "trtinterpret": [54, 65, 72], "along": [54, 76], "match": [54, 55, 93, 94, 102], "special": [54, 56, 104, 111], "account": [54, 114, 115, 117], "illustr": [54, 65, 104, 105, 109, 110, 111, 116], "scale_grad_by_freq": [54, 70], "embedding_param_valid": 54, "establish": 54, "subset": [54, 64, 71, 76, 91, 116], "converter_util": [54, 96], "enforce_tensor_typ": 54, "dictionari": [54, 76, 77, 92, 103], "between": [54, 55, 56, 60, 66, 76, 82, 83, 91, 93, 98, 100, 109, 113], "possibl": [54, 66, 82, 93, 94, 96, 98, 114, 115, 116, 117], "prefer": [54, 64, 66, 89], "keyword": [54, 62, 71, 75, 76, 77, 94, 103, 107], "both": [54, 56, 64, 66, 69, 71, 72, 75, 76, 80, 82, 88, 91, 94, 96, 98, 114, 115, 117], "enforc": [54, 89], "situat": 54, "partit": [54, 55, 63, 64, 71, 76, 94, 122], "greater": [54, 71, 73, 76], "than": [54, 55, 64, 66, 71, 76, 81, 82, 94, 97, 98, 100, 111, 113, 116, 120], "3d": [54, 65], "autocast": 54, "therebi": [54, 58, 93, 96, 116], "limit": [54, 55, 73, 81, 91, 95, 98, 99, 113, 114, 122], "author": [54, 83], "conv_nod": 54, "7": [54, 56, 58, 59, 75, 76, 86, 89, 93, 96, 103, 105, 107, 112, 118], "ignor": [54, 64, 71, 75, 76, 93, 96], "misc": [54, 93, 96], "trttensor": 54, "np": [54, 94, 96, 97, 98, 100, 101, 102, 111, 113, 114, 115, 117], "ndarrai": [54, 96], "aten_ops_convolut": 54, "conversioncontext": [54, 94, 96], "side": [54, 55, 80, 89, 94], "effect": [54, 55, 64, 65, 71, 80, 89, 91, 94, 96, 101, 116], "term": [54, 76, 82, 83, 91, 93, 94, 96, 116], "getitem": 54, "categor": 54, "modif": [54, 62, 76, 111], "op_evalu": 54, "capbility_valid": 54, "opcod": 54, "decompos": 54, "suboper": 54, "separ": [54, 56, 57, 59, 66], "Such": 54, "via": [54, 64, 65, 67, 69, 71, 75, 76, 77, 80, 86, 90, 91, 103, 105, 107, 109, 110, 111, 116, 118, 120, 121, 122], "register_torch_trt_decomposit": 54, "addmm_replac": 54, "replac": [54, 56, 62, 66, 67, 74, 93, 96, 112, 122], "input_": 54, "mat1": 54, "mat2": [54, 70], "beta": [54, 65, 70, 77], "alpha": [54, 65, 70, 83], "mul": [54, 56, 70, 94, 108], "matmul": [54, 55, 64, 70, 71, 89, 109, 110, 111, 118], "modifi": [54, 56, 62, 65, 83, 100, 118], "edit": [54, 66, 80], "torch_enabled_decomposit": 54, "torch_disabled_decomposit": 54, "disjoint": 54, "preced": [54, 82], "over": [54, 57, 59, 65, 82, 112, 113, 114, 115, 117, 122], "much": [54, 60, 80, 82, 91], "significantli": [54, 55, 80, 93, 98], "easier": [54, 57, 59, 60, 65, 71, 75, 76, 89, 91, 96, 100], "tri": 54, "made": [55, 57, 59, 76, 82], "represent": [55, 60, 65, 88, 104, 111, 116, 122], "instanc": [55, 62, 64, 66, 71, 74, 75, 88, 89, 94, 116, 120], "idea": [55, 82, 94], "reduc": [55, 56, 57, 59, 65, 71, 76, 91, 93, 96, 98, 101, 108, 116, 120], "actual": [55, 58, 60, 65, 88, 89, 96], "aim": [55, 122], "closer": 55, "scope": [55, 96, 103, 107], "csrc": [55, 61], "common_subexpression_elimin": 55, "subexpress": 55, "dead_code_elimin": 55, "exception_elimin": 55, "wa": [55, 58, 62, 64, 65, 71, 75, 76, 82, 89, 94, 95, 122], "1013": 55, "ne": [55, 70], "1012": 55, "24": [55, 67, 114, 115, 117], "lib": [55, 66, 67, 89], "python3": [55, 66, 89], "site": [55, 66, 82, 89], "nn": [55, 61, 65, 71, 72, 75, 76, 77, 88, 89, 90, 93, 94, 96, 103, 108, 111, 112, 118, 119, 122], "batchnorm": 55, "248": 55, "11": [55, 66, 82, 86, 89, 114, 115, 117], "block0": 55, "raiseexcept": 55, "249": 55, "12": [55, 56, 67, 82, 86, 88, 89, 105, 114, 115, 117, 118], "block1": 55, "guard_elimin": 55, "whose": [55, 65, 105], "freeze_modul": 55, "propag": 55, "fuse_addmm_branch": 55, "variant": [55, 120], "caught": 55, "ret": 55, "622": 55, "self": [55, 58, 60, 70, 75, 76, 88, 89, 90, 93, 94, 96, 98, 103, 108, 111, 112, 116, 118, 119, 123], "bia": [55, 70, 89, 112], "x9": 55, "3677": 55, "output0": [55, 114, 115, 117, 119], "add_": [55, 70, 89, 94], "fuse_linear": 55, "back": [55, 56, 58, 59, 75, 76, 82, 88, 89, 93, 96, 122], "fuse_flatten_linear": 55, "implicitli": [55, 76], "connect": [55, 71, 76, 77, 82, 100, 114, 115, 117, 123], "higher": [55, 64, 71, 76, 80, 82, 88, 113], "1d": 55, "lower_graph": 55, "access": [55, 60, 65, 80, 89, 92, 122], "rather": [55, 111], "getattr": [55, 58, 88, 89], "trainabl": 55, "remain": [55, 76, 91, 122], "lower_tupl": 55, "lowersimpletupl": 55, "tupleconstruct": [55, 58], "tupleunpack": 55, "leav": [55, 62, 64, 71], "statement": [55, 82, 94], "loweralltupl": 55, "_all_": 55, "rais": [55, 65, 76, 95], "onnx": 55, "module_fallback": 55, "consist": [55, 65, 82, 96, 101, 108, 111, 120, 122], "pair": [55, 60, 66, 82, 91, 116], "delimit": 55, "around": [55, 58, 60, 64, 66, 71, 75, 82, 85, 88, 96, 101], "second": [55, 65, 82, 90, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113], "mark": [55, 56, 80, 98, 104], "notatemoduleforfallback": 55, "marknodesforfallback": 55, "tell": [55, 56, 57, 58, 59, 60, 82, 93, 122], "them": [55, 56, 58, 63, 64, 65, 66, 71, 75, 80, 89, 93, 96, 98, 110, 111, 116, 118, 122], "peephole_optimz": 55, "intent": [55, 82], "catch": [55, 76, 89], "small": [55, 96, 97, 101, 114, 115, 117], "might": [55, 66, 80, 102, 118], "interest": [55, 82], "now": [55, 56, 59, 60, 65, 66, 76, 82, 89, 92, 93, 94, 96, 98, 102, 113, 119, 120], "expand": [55, 70], "simpli": [55, 103, 116], "remove_contigu": 55, "remove_dropout": 55, "infer": [55, 64, 65, 71, 76, 77, 89, 91, 95, 102, 103, 113, 116, 118, 120, 121, 122], "remove_to": 55, "unpack_addmm": 55, "reus": [55, 65, 71, 91, 93, 98, 101], "dedic": [55, 83], "unpack_log_softmax": 55, "softmax": [55, 65, 70, 112], "loop_unrol": 55, "suffici": [55, 66, 76], "short": [55, 64, 71, 82, 83, 102], "tile_to_repeat": 55, "instruct": [56, 57, 59, 65, 66, 89, 111, 114, 115, 117], "criteria": [56, 57, 59, 64], "lack": [56, 57, 59, 65, 93, 96, 113], "explicitli": [56, 57, 59, 66, 77, 90, 91, 92, 109, 110, 119], "On": 56, "segment": [56, 63, 96, 105, 107, 111, 116], "verifi": [56, 71, 94, 96, 102], "Then": [56, 91, 92, 102, 114, 115, 117], "roughli": [56, 114, 115, 117], "analysi": 56, "everi": [56, 72, 75, 76, 89, 120], "complet": [56, 63, 71, 76, 88, 89], "mean": [56, 60, 64, 65, 70, 71, 72, 103, 113, 114, 115, 117, 122], "trace": [56, 65, 71, 75, 77, 88, 89, 114, 115, 117, 118, 121, 122], "tensorlist": [56, 60], "figur": [56, 83, 85, 111], "our": [56, 59, 63, 88, 89, 114, 115, 117], "stitch": [56, 89], "altogeth": [56, 80], "brief": 56, "descript": [56, 83, 95, 112], "partitioninfo": 56, "api": [56, 59, 60, 62, 63, 64, 65, 75, 76, 77, 81, 89, 90, 91, 92, 96, 103, 104, 105, 108, 113, 114, 115, 116, 117, 118, 120, 121], "maintain": [56, 58, 60, 76, 100, 108, 122], "code": [56, 59, 62, 64, 65, 66, 81, 83, 88, 89, 91, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 118], "mymodel": [56, 63, 68, 90, 93, 96, 118, 121], "ts_model": [56, 89], "trt_model": [56, 92, 96, 105, 109, 110, 111, 112, 113, 114, 115, 117, 121], "off": [56, 58, 108, 111], "consecut": [56, 63], "satisfi": [56, 62, 65], "forced_fallback_op": 56, "randn": [56, 63, 68, 71, 76, 77, 89, 92, 94, 98, 105, 108, 118, 119, 121], "224": [56, 63, 68, 71, 72, 76, 77, 89, 95, 98, 100, 102, 105, 108, 114, 115, 116, 117, 118, 121], "trt_ts_modul": [56, 90], "input_s": 56, "inputrang": 56, "cfg": [56, 89], "relu": [56, 70, 88, 89, 103, 108, 112], "trt_mod": [56, 68, 89, 91, 123], "consid": [56, 77, 89, 96, 119], "segmentmodelwithdependencyawar": 56, "test_segment": 56, "20": [56, 67, 86, 102, 105, 107], "x_lgamma": 56, "lgamma": 56, "y_lgamma": 56, "div": [56, 70], "div_lgamma": 56, "27": [56, 89], "cat": [56, 66, 67, 70, 112, 113], "greedi": [56, 104, 109, 110, 113], "strategi": [56, 76], "travers": [56, 57, 59, 64], "gather": 56, "same": [56, 58, 62, 64, 65, 66, 71, 76, 80, 82, 88, 89, 92, 93, 95, 96, 98, 102, 105, 107, 114, 115, 117, 118, 120, 121], "encount": [56, 64, 66, 94, 103, 104, 107], "4": [56, 58, 63, 64, 65, 66, 70, 76, 78, 80, 82, 83, 86, 89, 96, 103, 106, 107, 108, 112, 118], "suboptim": 56, "arithmet": 56, "split": [56, 65, 70], "own": [56, 60, 64, 66, 71, 82, 89, 98, 112, 114, 115, 117], "could": [56, 64, 65, 96, 105, 107, 120], "rewrit": [56, 62], "portion": [56, 82, 96, 106], "without": [56, 60, 68, 71, 80, 82, 89, 91, 96, 97, 98, 102, 120], "reorder": 56, "seri": 56, "cleanli": 56, "approach": [56, 98], "achiev": [56, 116], "hit": 56, "larger": [56, 71, 76, 80, 113, 116], "boundari": [56, 74, 76], "guarante": [56, 75], "trigger": [56, 64, 65, 76, 89, 98, 100, 102, 122], "appear": [56, 82], "adjac": [56, 71, 76, 82], "As": [56, 65, 66, 76, 89, 93, 94, 96, 98, 102, 108, 122], "clean": [56, 62, 82, 103, 107], "step": [56, 65, 67, 70, 76, 91, 96, 102, 111, 116], "consolid": [56, 88], "further": [56, 64, 65, 120, 122], "merg": 56, "identifi": 56, "do_not_merg": 56, "combin": [56, 64, 65], "condit": [56, 82, 122], "loop": [56, 64, 65, 104, 109, 110], "ir": [57, 59, 60, 63, 64, 68, 71, 76, 88, 89, 90, 99, 101, 103, 105, 107, 108, 114, 118], "larg": [57, 59, 80, 82, 89, 91, 101, 102, 104, 111, 113, 116], "opset": [57, 59, 94], "compon": [57, 59, 66, 67, 74, 88, 120, 122], "evalu": [57, 58, 59, 112], "deploi": [57, 59, 69, 71, 89, 91, 99, 114, 115, 117], "instanti": [57, 58, 59, 60, 89, 106], "wrap": [57, 58, 59, 65, 82, 85, 89, 92, 103, 107, 108], "extend": [57, 59, 60, 70, 89, 98, 116], "providi": [57, 59], "stand": [58, 82], "interpret": [58, 65, 82], "execute_engin": [58, 75, 89], "stack": [58, 70, 91, 112, 122], "machin": [58, 66, 91, 95, 114, 115, 117], "pop": 58, "push": 58, "element": [58, 65, 82, 83, 86, 93], "realiz": 58, "abstract": [58, 60, 83, 94], "__torch__": [58, 88, 89], "portabl": [58, 66, 77], "serializ": [58, 64, 88, 122], "instnanti": 58, "whatev": [58, 65, 96], "self_1": [58, 89], "torchvis": [58, 91, 92, 95, 98, 100, 102, 105, 108, 112, 114, 115, 117], "resnet": [58, 69, 78, 95, 99, 100, 114, 115, 116, 117], "___torch_mangle_4847": 58, "resnet_trt": 58, "input_0": [58, 89], "__torch___torchvision_models_resnet____torch_mangle_4847_resnet_trt_engin": 58, "listunpack": [58, 89], "multipl": [58, 66, 71, 75, 76, 82, 83, 91, 93, 101, 113, 114, 115, 117, 120], "repack": 58, "ssd": 58, "ssd300_trt": 58, "__torch___pytorch_detection_ssd_src_model_ssd300_trt_engin": 58, "holder": [58, 84], "torchbind": 58, "pickler": 58, "seril": 58, "zip": [58, 66, 100, 102, 111, 114], "depickl": 58, "encod": [58, 111, 116], "sm": 58, "correct": [58, 66, 80, 99, 100, 102, 112, 114, 115, 117], "bazel": [59, 66, 67], "linux": [59, 64, 67, 71, 89, 95], "x86_64": [59, 66], "aarch64": 59, "gcc": [59, 89], "untest": 59, "try": [59, 76, 82, 83, 89, 92, 96, 98, 111, 114, 115, 117, 122], "older": 59, "repositori": [59, 66, 80, 87, 111, 114, 115, 117], "notebook": [59, 69, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114], "doc": [59, 61, 66, 67, 80, 81, 82, 87, 94, 96, 118], "docsrc": 59, "third_parti": [59, 66], "toolchain": [59, 66, 67], "unstabl": 59, "subject": [59, 62, 122], "matur": 59, "most": [59, 65, 66, 72, 96, 102, 114, 115, 117, 120, 122], "hood": [59, 105, 122], "major": [59, 65, 76], "top": [59, 80, 84], "coordin": [59, 76, 111], "ingest": 59, "flow": [60, 65, 82, 88, 116], "ilay": 60, "analogu": 60, "goal": [60, 64, 98], "registernodeconversionpattern": [60, 89], "helper": [60, 94], "pattern": [60, 76, 89, 113], "schema": [60, 89, 94, 96], "caus": [60, 64, 80, 103, 105, 107, 113, 120], "acthardtanh": 60, "torchtrt_unus": 60, "hardtanh": [60, 70], "scalar": [60, 70], "min_val": [60, 70], "max_val": [60, 70], "unwraptodoubl": 60, "new_lay": 60, "addactiv": 60, "activationtyp": [60, 65], "kclip": 60, "torchtrt_check": 60, "unabl": [60, 89, 96], "setalpha": 60, "setbeta": 60, "setnam": [60, 89], "util": [60, 62, 74, 77, 89, 91, 103, 107, 109, 110, 111, 112, 113, 114, 115, 116, 117, 122], "node_info": [60, 89], "c_str": [60, 89], "out_tensor": [60, 89], "associatevalueandtensor": [60, 89], "getoutput": [60, 89], "log_debug": 60, "getdimens": [60, 89], "accord": [60, 64, 77, 95], "unwrap": 60, "tool": [60, 64, 65, 66, 89, 94, 98, 116], "don": [60, 65, 80, 82, 83, 91, 94, 112, 114, 115, 117, 118], "annot": [60, 89], "your": [60, 63, 64, 66, 67, 68, 75, 80, 82, 83, 87, 88, 89, 90, 92, 98, 111, 118, 120], "Its": [60, 82], "track": [60, 91], "sort": [60, 70, 92, 111], "live": [60, 82], "directli": [60, 62, 63, 66, 69, 74, 76, 91, 94, 96, 103, 111, 121], "associatevalueandivalu": 60, "inspect": [60, 88, 89], "dataflow": [60, 89], "mechan": [60, 64, 65, 96, 101, 102, 116], "safe": [60, 64, 71, 75, 76, 77, 111], "unsur": 60, "deep": [60, 64, 69, 80, 91, 96, 123], "straight": 60, "chanc": 60, "none": [60, 64, 65, 70, 71, 72, 74, 75, 76, 77, 80, 82, 94, 96, 98, 103, 104, 111, 112, 113], "wrapper": [60, 65, 101, 108, 121], "similar": [60, 63, 64, 65, 66, 89, 92, 93, 96, 109, 110, 111], "tocustomclass": 60, "tensorcontain": 60, "istensor": 60, "iscustomclass": 60, "lot": [60, 63], "singular": 60, "becaus": [60, 65, 66, 72, 88, 89, 93, 94, 96, 97, 98, 101, 108, 113, 119], "alloc": [60, 69, 99, 108, 114], "freed": 60, "destructor": 60, "destroi": [60, 83], "realli": 60, "think": [60, 82], "becom": [60, 66, 100], "benefit": [60, 89, 98, 108, 113], "deal": [60, 98], "quit": [60, 66, 89, 116], "effici": [60, 101, 108, 111], "batch_norm": [60, 70], "fusion": [60, 62, 65], "deeplearn": [61, 65, 67], "sdk": [61, 67, 114, 115, 117, 122], "matrix": 61, "html": [61, 66, 67, 82, 88, 91, 94, 96, 118], "c_api": 61, "python_api": 61, "org": [61, 66, 80, 82, 88, 89, 91, 94, 96, 118, 120], "stabl": [61, 67, 69, 77, 78, 80, 99, 114, 118], "master": [61, 66, 91, 120], "overview": [61, 69, 103, 108], "md": 61, "appli": [62, 63, 91, 102, 104, 108, 111], "desir": [62, 71, 83, 91, 98], "coalesc": 62, "insert": [62, 64, 71, 89, 91, 94, 98, 102], "graphmodul": [62, 63, 71, 72, 76, 89, 90, 96, 102, 121, 122], "caller": 62, "invok": [62, 64, 65, 88, 89, 120], "lint": 62, "recompil": [62, 71, 76, 94, 98, 102, 104, 107, 118, 122], "repair": 62, "disallow": 62, "repair_input_as_output": 62, "gm": [62, 71], "sample_input": [62, 65, 103], "scenario": [62, 64, 100, 101, 113], "clone": [62, 66, 70, 96], "modified_graph": 62, "extract": [62, 89, 111, 116], "placehold": [62, 94], "isinst": [62, 65, 96, 112], "issubclass": 62, "direct": [62, 86, 102, 120], "len": [62, 70, 96, 111], "direct_output": 62, "inserting_aft": 62, "cloned_placehold": 62, "replace_input_with": 62, "date": [62, 83, 122], "eliminate_dead_cod": 62, "logger": [62, 73], "f": [62, 64, 65, 67, 76, 82, 88, 94, 95, 96, 101, 111, 112, 113], "__init__": [62, 75, 76, 82, 88, 93, 94, 96, 98, 103, 111, 112, 118, 119], "pass_manag": 62, "passmanag": 62, "backend": [62, 68, 69, 77, 78, 81, 92, 97, 98, 99, 103, 104, 108, 112, 114, 115, 117, 118], "offer": [62, 64], "registr": [62, 65], "conveni": [62, 91, 107, 116, 120, 122], "control": [62, 65, 88, 102, 113, 120], "_aten_lowering_pass": 62, "my_custom_pass": 62, "front": [62, 71], "passlist": 62, "arbitrari": [62, 75], "remov": [62, 63, 71, 80, 97, 98, 101, 111, 112], "dump_lowering_pass": 62, "apply_lowering_pass": 62, "graph_modul": [62, 71], "_remove_lowering_pass": 62, "evolv": 62, "introduc": [63, 65, 108, 116], "exportedprogram": [63, 68, 71, 76, 102, 109, 110, 113, 118, 122], "dynamo": [63, 64, 66, 68, 74, 75, 76, 78, 89, 93, 94, 95, 96, 97, 98, 99, 101, 102, 103, 105, 107, 108, 112, 113, 114, 115, 117, 118, 119], "frontend": [63, 71, 74, 90, 93, 96, 99, 105, 107, 112, 114, 115, 117, 118], "simpl": [63, 64, 65, 82, 83, 88, 93, 114, 115, 116, 117, 118], "usag": [63, 65, 69, 74, 78, 82, 89, 93, 99, 113, 114, 118, 121], "eval": [63, 68, 89, 90, 94, 95, 97, 98, 100, 101, 102, 103, 104, 105, 107, 108, 109, 110, 111, 112, 113, 114, 115, 117, 118, 119, 121], "exp_program": [63, 98, 102, 111, 112, 118], "trt_gm": [63, 68, 98, 102, 118, 119, 121], "interact": [63, 82, 100, 103, 105, 106, 107, 108], "ideal": 63, "discuss": [63, 64, 114, 115, 117], "section": [63, 65, 80, 82, 83, 84, 86, 89, 91, 114, 115, 117, 121], "frequent": [63, 101], "builder": [63, 64, 65, 71], "respect": [63, 64, 66, 71, 76, 109, 110, 119], "releas": [63, 64, 67, 82], "insid": [63, 82, 93, 96], "decomposit": [63, 64, 71, 76, 93, 96], "downstream": [63, 116], "constraint": [63, 113], "guid": [64, 81], "present": [64, 102], "learn": [64, 66, 69, 89, 91, 96, 114, 115, 117, 123], "acceler": [64, 72, 76, 120, 122, 123], "workflow": [64, 65, 68, 69, 71, 72, 76, 89, 92, 98, 99, 100, 105, 106, 107, 109, 110, 114, 116], "wide": [64, 76, 86], "varieti": [64, 114, 115, 117], "primari": [64, 94, 98, 121], "simplic": 64, "optimized_model": [64, 68, 97, 101, 103, 105, 107], "depth": [64, 80, 116], "challeng": [64, 100, 114, 115, 117], "addition": [64, 96], "fit": [64, 82], "compilationset": [64, 71, 75, 94, 96, 103], "_enum": [64, 71], "callabl": [64, 71, 76], "pass_through_build_failur": [64, 71, 75, 76, 96, 108], "max_aux_stream": [64, 71, 75, 76, 96], "version_compat": [64, 71, 75, 76, 96], "optimization_level": [64, 71, 75, 76, 96, 103], "use_python_runtim": [64, 71, 75, 76, 96, 97, 98, 100, 102, 103], "truncate_doubl": [64, 71, 75, 76, 96, 97, 109, 110, 113], "use_fast_partition": [64, 71, 75, 76, 96], "enable_experimental_decomposit": [64, 71, 75, 76, 96], "_devic": [64, 71], "assume_dynamic_shape_support": [64, 71, 75, 76], "engine_cap": [64, 71, 75, 76, 96], "dryrun": [64, 71, 75, 76, 96], "hardware_compat": [64, 71, 75, 76, 96], "timing_cache_path": [64, 71, 75, 76, 98], "tmp": [64, 71, 75, 76, 89, 97], "torch_tensorrt_engine_cach": [64, 71, 75, 76], "timing_cach": [64, 65, 71, 75, 76], "bin": [64, 66, 67, 71, 75, 76], "lazy_engine_init": [64, 71, 75, 76], "cache_built_engin": [64, 71, 75, 97, 98], "reuse_cached_engin": [64, 71, 75, 97, 98, 102], "use_explicit_typ": [64, 71, 75, 109, 110, 113, 119], "use_fp32_acc": [64, 71, 75, 109, 110, 111], "refit_identical_engine_weight": [64, 71, 75], "strip_engine_weight": [64, 71, 75], "immutable_weight": [64, 71, 75, 76, 97, 98, 100, 102], "enable_weight_stream": [64, 71, 75, 113], "enable_cross_compile_for_window": [64, 71, 75], "dpython": [64, 71, 76, 77], "per": [64, 71, 96, 120], "regardless": [64, 71, 83, 105, 107], "fail": [64, 71, 76, 89, 100, 102, 112, 123], "auxiliari": [64, 71], "stream": [64, 69, 71, 76, 93, 96, 99, 114], "impli": [64, 71], "longer": [64, 66, 71, 76, 80, 95, 120], "search": [64, 69, 71, 76, 80], "strictli": [64, 71], "runtim": [64, 66, 68, 69, 71, 76, 89, 94, 99, 100, 103, 107, 108, 113, 114, 122], "presenc": [64, 71, 108], "preferenti": [64, 71], "choos": [64, 65, 71, 88], "float64": [64, 71, 76, 77], "toggl": [64, 71, 76], "mode": [64, 65, 71, 75, 76, 90, 91, 94, 108, 111, 112], "detail": [64, 65, 67, 71, 88, 89, 96, 98, 114, 115, 117, 120], "natur": [64, 71, 82], "architectur": [64, 66, 69, 71, 76, 95, 98, 116], "amper": [64, 71, 76], "newer": [64, 66, 71, 76], "storag": [64, 71, 91], "use_strong_typ": [64, 71], "strong": [64, 71, 82], "mix": [64, 69, 71], "happen": [64, 65, 71, 88, 100, 105, 118], "strip": [64, 71], "non": [64, 66, 71, 76, 83, 85, 111, 120], "refitt": [64, 71, 76, 98], "were": [64, 71, 96, 102, 120], "cross": [64, 71, 82, 99, 114], "window": [64, 71, 82], "sub": [64, 70, 82, 88, 103], "slate": 64, "futur": [64, 65, 71, 76, 77, 104, 120], "occur": [64, 108, 113], "first_output": 64, "subsequ": [64, 98, 101, 108], "second_output": 64, "session": [64, 68, 82, 98, 108], "point": [64, 66, 76, 80, 81, 82, 89, 93, 111, 112, 114, 115, 117], "cover": [64, 93, 94], "benchmark": [64, 70], "automat": [64, 67, 76, 82, 89, 99, 102, 114, 118, 122], "vari": [64, 72, 113, 118], "distribut": [64, 67, 89, 91, 113, 120], "inf": 64, "dynamo_convers": 64, "contribut": [64, 101], "demonstr": [64, 82, 83, 84, 91, 93, 94, 96, 98, 100, 112, 114, 115, 116, 117], "break": [64, 65, 71, 75, 76, 82, 93, 96, 101, 110, 111], "successfulli": [64, 95, 100, 102, 111], "_dynamo": [64, 97, 98, 103, 104, 105, 107, 118], "explain": [64, 65, 69], "veri": [64, 65, 83, 84, 91, 92, 104, 109, 110, 114, 115, 117], "explan": [64, 65], "graph_break_count": 64, "furthermor": 64, "durat": [64, 82], "latter": [64, 75], "logic": [64, 65, 94], "guard": 64, "compos": [65, 88, 91, 94, 112, 114, 115, 117], "variou": [65, 123], "etc": [65, 80, 82, 96, 123], "environ": [65, 68, 71, 114, 115, 117], "research": 65, "few": [65, 66, 76, 94], "nightli": 65, "lower_exampl": 65, "welcom": [65, 89], "finish": 65, "converison": 65, "pleas": [65, 67, 76, 82, 89, 99, 111, 112, 114, 115, 117, 118], "max_batch_s": [65, 72, 114, 115, 117], "2048": [65, 72], "max_workspace_s": [65, 72], "33554432": [65, 72], "explicit_batch_dimens": [65, 72], "lower_precis": [65, 72], "lowerprecis": [65, 72], "verbose_log": [65, 72], "timing_cache_prefix": [65, 72], "save_timing_cach": [65, 72], "cuda_graph_batch_s": [65, 72], "dynamic_batch": [65, 72], "turn": [65, 72, 108], "trtmodul": [65, 72], "otherwis": [65, 66, 72, 98, 120], "implicit": [65, 70, 72, 82], "config": [65, 66, 72, 114, 115, 117], "updat": [65, 66, 67, 71, 72, 76, 96, 102], "dim": [65, 70, 72, 96, 98, 112, 113, 114, 115, 117, 118], "fx2trt_exampl": 65, "acc_trac": 65, "come": [65, 66, 81, 93, 96, 100, 114, 115, 117], "my_pytorch_model": 65, "build_model": 65, "prepar": [65, 114, 115, 117], "acc_mod": 65, "earli": [65, 102], "deprec": [65, 70], "continu": [65, 82, 108, 120], "backward": [65, 75, 96, 122], "vision": [65, 99, 114, 115, 117], "activ": [65, 77, 82, 89, 91, 94, 116, 120, 123], "except": [65, 71, 76], "permut": [65, 70, 111], "transpos": [65, 70, 118], "ll": [65, 98, 104], "inputtensorspec": [65, 72, 76], "experiment": [65, 76, 77], "dataclass": [65, 103], "re": [65, 76, 82, 93, 98, 100, 108, 120], "manual": [65, 76, 81, 82, 102, 113], "sampl": [65, 71, 82, 90, 91, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 114, 115, 117], "rand": [65, 89, 95, 98, 100, 102, 103, 114, 115, 117], "from_tensor": [65, 76], "slightli": [65, 66, 96], "promis": 65, "optimize_target_shap": 65, "input_tensor_spec": 65, "shape_rang": [65, 72], "100": [65, 72, 96, 98, 112, 113], "accordingli": [65, 80, 118, 120], "trtinterpreterresult": [65, 72], "namedtupl": 65, "input_nam": [65, 72], "output_nam": [65, 72], "serialized_cach": [65, 72], "bytearrai": [65, 75, 77], "afford": 65, "temporari": [65, 98], "best": [65, 71, 76, 82, 100, 113, 119], "perforamnc": 65, "examin": 65, "suitabl": [65, 94, 101], "force_fp32_output": 65, "strict_type_constraint": 65, "usual": [65, 66, 80, 104], "unless": 65, "certain": [65, 66, 103, 109, 110, 111, 113, 120], "algorithm_selector": 65, "profiling_verbos": 65, "trt_interpreter_result": 65, "64": [65, 76, 90, 93, 110, 111, 112, 118], "25": [65, 72, 89, 111], "runtimeerror": [65, 112], "xxx": 65, "One": [65, 82, 83, 89, 116, 120], "reload_trt_mod": 65, "reload_model_output": 65, "far": [65, 82], "give": [65, 80, 82], "convtert": 65, "scheme": [65, 71, 76], "action": [65, 82], "tensort": [65, 122], "thing": [65, 66, 82], "compar": [65, 71, 76, 90, 101, 102], "vanilla": 65, "mainli": 65, "builtin": 65, "purpos": [65, 114, 115, 116, 117], "acc_op": 65, "leverag": [65, 91], "power": [65, 82, 89, 113, 116], "goe": [65, 82], "whole": [65, 108], "sigmoid": [65, 70], "tensorrt_convert": 65, "acc_ops_sigmoid": 65, "rest": [65, 82, 83], "input_v": [65, 94], "receiv": 65, "region": 65, "add_activ": 65, "get_output": [65, 96], "wherev": 65, "rememb": [65, 66, 114, 115, 117], "mapper": 65, "todo": [65, 67, 80], "logist": 65, "down": [65, 66, 80, 110], "acc_norm": 65, "foo": [65, 82, 83], "register_acc_op": 65, "register_acc_op_map": 65, "this_arg_is_opt": 65, "op_and_target": 65, "arg_replacement_tupl": 65, "rule": [65, 66, 77], "third": [65, 83], "boolean": [65, 76, 94], "matter": [65, 96], "register_custom_acc_mapper_fn": 65, "design": [65, 74, 94, 100, 109, 113, 116, 123], "redund": 65, "throught": 65, "custom_mapp": 65, "_": [65, 82, 93, 96, 101, 111, 112, 113, 119], "foo_kwarg": 65, "inserting_befor": 65, "foo_nod": 65, "meta": [65, 67, 86, 93, 110, 113], "children": 65, "unit": [65, 76, 108], "test_acc_trac": 65, "acc_op_convert": 65, "essenti": 65, "plugin": [65, 96], "yet": [65, 116], "folder": 65, "center": 66, "pypi": 66, "m": [66, 67, 83, 93, 104, 112], "pip": [66, 67, 99, 104, 111, 114, 115, 117], "upload": [66, 114, 115, 117], "x86": [66, 120], "extra": [66, 75, 89, 96, 100], "url": [66, 80, 114, 115, 117], "download": [66, 67, 86, 91, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 117], "whl": [66, 67], "cu118": 66, "cu124": 66, "tarbal": [66, 89, 91], "easiest": [66, 93, 96], "bazelisk": [66, 67], "bazelbuild": [66, 67], "export": [66, 67, 69, 71, 76, 98, 99, 102, 105, 109, 110, 111, 112, 113, 114, 115, 117, 119, 121, 122], "bazel_vers": 66, "path_to_torchtrt_root": 66, "bazelvers": 66, "mkdir": 66, "cd": [66, 114, 115, 117], "curl": [66, 82], "fssl": 66, "o": [66, 82, 114, 115, 117], "dist": 66, "unzip": 66, "bash": [66, 114, 115, 117], "sh": 66, "cp": [66, 67, 96], "usr": [66, 67], "driver": 66, "branch": [66, 67], "4e5b0f6e860910eb510fa70a76ee3eb9825e7a4d": 66, "l46": 66, "pull": [66, 98, 114, 115, 117], "latest": [66, 67, 80], "l53c1": 66, "fact": 66, "reproduc": 66, "l71": 66, "http_archiv": 66, "build_fil": 66, "archiv": [66, 67], "sha256": 66, "strip_prefix": 66, "OR": 66, "TO": [66, 89], "gnu": 66, "tar": [66, 67, 82, 91], "gz": [66, 82, 83, 91], "ld_library_path": 66, "comment": [66, 82], "uncom": 66, "l114c1": 66, "l124c3": 66, "uv": 66, "astral": 66, "project": [66, 81, 86], "simpler": [66, 91], "wheel": [66, 67], "dep": 66, "lighter": 66, "executor": 66, "avoid": [66, 93, 94, 96, 102, 111, 118], "implic": 66, "python_onli": 66, "legaci": [66, 74], "mainten": 66, "torchdynamo": [66, 118, 122], "technolog": [66, 122], "exclud": [66, 96], "speed": [66, 98, 102], "no_torchscript": 66, "dbg": 66, "pre_cxx11_abi": 66, "complic": 66, "incompat": 66, "popular": [66, 81, 99, 109, 110, 114, 116], "ngc": [66, 67, 114, 115, 117], "tabl": [66, 86], "bdist_wheel": 66, "preinstal": 66, "forum": 66, "correctli": [66, 96], "declar": 66, "intend": [66, 103, 105, 106, 107, 108], "microsoft": 66, "2022": [66, 69], "open": [66, 111, 114, 115, 116, 117], "app": 66, "x64": 66, "prompt": [66, 100, 104, 106, 109, 110, 111], "admin": 66, "privileg": 66, "launcher": 66, "chocolatei": 66, "navig": [66, 80], "ninja": 66, "setuptool": 66, "r": [66, 67, 82, 99, 104, 111, 114], "txt": [66, 67, 99, 104, 111, 114], "distutils_use_sdk": 66, "cuda_win": 66, "libtorch_win": 66, "tensorrt_win": 66, "similarli": [66, 98, 108, 120], "ci_workspac": 66, "win": 66, "tmpl": [66, 67], "torchtrtc": [66, 69, 123], "websit": 66, "finder": 66, "dcmake_module_path": 66, "doesn": [66, 82, 88, 89], "dtorch_dir": 66, "dtensorrt_root": 66, "choic": [66, 74], "b": [66, 70, 76, 83, 93, 113, 114, 115, 117], "dcmake_build_typ": 66, "72048": 66, "jp_workspac": [66, 67], "new_local_repositori": 66, "sudo": [66, 67], "home": 66, "unlik": [66, 92], "libtorch_pre_cxx11_abi": 66, "shift": [66, 70, 82], "jetpack": 66, "jetpack_x": 66, "jetpack_5": 66, "drop": [66, 80, 112], "nvida": 67, "ofjetpack": 67, "With": [67, 80, 82, 89, 91, 93, 98, 114, 115, 117], "incorpor": [67, 83], "cudnn": 67, "9": [67, 86, 89, 95, 96, 114, 115, 117], "dlfw": 67, "09": 67, "jetson": [67, 116], "framework": 67, "instal": [67, 69, 86, 89, 99, 104, 111, 114, 115, 117, 120], "kit": 67, "flash": 67, "board": 67, "apt": 67, "show": [67, 80, 82, 98, 106, 113, 116], "dev": 67, "everth": 67, "nvcc": 67, "cmd": 67, "toolkit": [67, 74], "libcusparselt": 67, "lib64": 67, "wget": [67, 114, 115, 117], "cusparselt": 67, "redist": 67, "libcusparse_lt": 67, "sbsa": 67, "xz": 67, "xf": 67, "v1": [67, 100, 106], "arm64": 67, "mv": 67, "chmod": 67, "pypa": 67, "en": [67, 80], "bootstrap": 67, "jp": 67, "v61": 67, "0a0": 67, "872d972e41": 67, "nv24": 67, "08": [67, 114, 115, 117], "17622132": 67, "cp310": 67, "linux_aarch64": 67, "test_requir": 67, "jetpack6": 67, "lanl": 67, "cuda_vers": 67, "grep": 67, "cut": [67, 82, 102], "sed": [67, 83, 85], "torch_install_path": 67, "dirnam": 67, "__file__": 67, "site_package_path": 67, "cuda_hom": 67, "envsubst": 67, "cxx11": [67, 120], "abi": [67, 120], "anywher": 68, "ahead": [68, 69, 89, 100, 108], "ep": [68, 70, 95, 102, 119, 121], "output_format": [68, 76, 121], "input_tensor": [68, 96, 112, 113], "fill": 68, "aot": [69, 89, 99, 100, 102, 108, 114, 122], "integr": [69, 100, 103], "seamlessli": [69, 76], "ecosystem": [69, 122], "hybrid": [69, 71, 76, 77, 122], "advanc": [69, 78, 83, 91, 99, 104, 114], "bert": [69, 78, 99, 101, 114], "triton": [69, 93, 96], "cudagraph": [69, 99, 114], "overload": [69, 99, 114], "mutabl": [69, 99, 114], "diffus": [69, 78, 99, 114], "gpt2": [69, 99, 114], "llama2": [69, 99, 114], "sam2": [69, 99, 114], "page": [69, 84, 86, 114, 115, 117], "introductori": 69, "blog": [69, 120], "gtc": 69, "2020": [69, 89], "talk": 69, "fall": [69, 76, 93, 96], "2021": 69, "dai": 69, "confer": 69, "_convolut": [70, 89], "stride": [70, 76, 96, 112], "pad": [70, 76, 96, 112], "dilat": 70, "output_pad": 70, "group": [70, 82, 83], "determinist": 70, "cudnn_en": 70, "allow_tf32": 70, "ab": 70, "aco": 70, "acosh": 70, "adaptive_avg_pool1d": 70, "output_s": 70, "adaptive_avg_pool2d": 70, "adaptive_avg_pool3d": 70, "adaptive_max_pool1d": 70, "adaptive_max_pool2d": 70, "adaptive_max_pool3d": 70, "argmax": [70, 113], "keepdim": 70, "argmin": 70, "asin": 70, "asinh": 70, "atan": 70, "atanh": 70, "avg_pool1d": 70, "kernel_s": [70, 96, 112], "ceil_mod": 70, "count_include_pad": 70, "avg_pool2d": 70, "divisor_overrid": 70, "avg_pool3d": 70, "gamma": 70, "var": 70, "momentum": 70, "bitwise_not": 70, "bmm": 70, "ceil": 70, "clamp": 70, "clamp_max": 70, "clamp_min": 70, "constant_pad_nd": 70, "co": [70, 83, 116], "cosh": 70, "cumsum": 70, "tensor_mod": 70, "rounding_mod": 70, "div_": 70, "elu": 70, "scale": [70, 91, 116], "input_scal": 70, "indic": [70, 71, 80, 82, 93, 94, 102, 105, 118, 119], "padding_idx": 70, "eq": [70, 82], "erf": [70, 94], "exp": 70, "expand_a": 70, "fake_quantize_per_channel_affin": 70, "zero_point": 70, "axi": [70, 76, 111], "quant_min": 70, "quant_max": 70, "fake_quantize_per_tensor_affin": 70, "using_int": [70, 89], "start_dim": [70, 89], "end_dim": [70, 89], "floor": 70, "floor_divid": 70, "ge": 70, "gru_cel": 70, "hx": 70, "w_ih": 70, "w_hh": 70, "b_ih": 70, "b_hh": 70, "gt": 70, "hardtanh_": 70, "instance_norm": 70, "running_mean": 70, "running_var": 70, "use_input_stat": 70, "layer_norm": 70, "normalized_shap": 70, "le": 70, "negative_slop": 70, "01": [70, 83, 89, 111, 112], "leaky_relu_": 70, "lstm_cell": 70, "lt": 70, "masked_fil": 70, "mask": [70, 96, 111], "max_pool1d": 70, "max_pool2d": [70, 88, 89], "max_pool3d": 70, "mul_": [70, 94], "narrow": 70, "neg": [70, 100], "norm": 70, "scalaropt_dim": 70, "pixel_shuffl": 70, "upscale_factor": 70, "pow": 70, "tensor_scalar": 70, "expon": 70, "tensor_tensor": 70, "prelu": 70, "prod": [70, 96], "dim_int": 70, "reciproc": 70, "reflection_pad1d": 70, "reflection_pad2d": 70, "relu_": 70, "repeat_interleav": 70, "self_int": 70, "replication_pad1d": 70, "replication_pad2d": 70, "replication_pad3d": 70, "reshap": [70, 96, 111], "roll": 70, "rsub": 70, "scatter": [70, 111], "sigmoid_": 70, "sin": [70, 82], "sinh": 70, "slice": 70, "split_siz": 70, "split_with_s": 70, "sqrt": 70, "squar": 70, "squeez": [70, 111, 116], "sub_": 70, "dim_intlist": 70, "tan": 70, "tanh": [70, 94], "tanh_": [70, 94], "non_block": [70, 112], "memory_format": [70, 76], "prim_devic": 70, "topk": 70, "k": [70, 91, 112], "largest": 70, "dim0": [70, 98], "dim1": 70, "unbind": 70, "unsqueez": [70, 111, 114, 115, 117], "upsample_bilinear2d": 70, "align_corn": 70, "scales_h": 70, "scales_w": 70, "vec": 70, "scale_factor": 70, "upsample_linear1d": 70, "upsample_nearest1d": 70, "upsample_nearest2d": 70, "upsample_nearest3d": 70, "scales_d": 70, "upsample_trilinear3d": 70, "view": [70, 80, 111], "__and__": 70, "__derive_index": 70, "idx": 70, "__getitem__": 70, "__is__": 70, "t1": 70, "t2": 70, "obj": 70, "__isnot__": 70, "__not__": 70, "__or__": 70, "__range_length": 70, "lo": 70, "hi": [70, 82, 83], "__round_to_zero_floordiv": 70, "__xor__": 70, "append": [70, 94, 97, 98, 101, 112, 113], "el": 70, "arang": [70, 93, 96], "pin_memori": 70, "start_step": 70, "copy_": 70, "float_int": 70, "int_float": 70, "floordiv": 70, "is_floating_point": 70, "numel": [70, 93], "l": [70, 112], "9223372036854775807": 70, "requires_grad": 70, "tupleindex": 70, "tup": 70, "exported_program": [71, 76, 121], "arg_input": [71, 76, 94, 102], "kwarg_input": [71, 76, 102], "engine_cache_dir": [71, 97, 98], "engine_cache_s": [71, 97, 98], "5368709120": 71, "custom_engine_cach": [71, 98], "baseenginecach": [71, 98], "int32": [71, 76, 77, 96, 97, 101, 107, 116], "channel_last": [71, 76, 77, 116], "244": [71, 76, 77], "alia": [71, 76], "better": [71, 76, 88, 111, 116, 122], "understand": [71, 76, 118], "convolut": [71, 76, 77, 91, 96, 123], "_c": [71, 76, 77, 92], "oppos": [71, 76, 77], "lean": [71, 76], "spend": [71, 76], "integ": [71, 76, 85], "faster": [71, 76, 97, 98, 101, 111, 116], "parition": [71, 76], "increas": [71, 76, 98, 113], "amount": [71, 76, 113], "defer": [71, 76, 122], "lead": [71, 76, 82, 101, 113, 120], "oversubscript": [71, 76], "hard": [71, 102], "disk": [71, 76, 98], "space": [71, 82, 83, 91], "byte": [71, 75, 76, 77, 96, 98, 113, 116], "1gb": [71, 97, 98], "exce": 71, "oldest": 71, "gear": [71, 91], "toward": [71, 91, 111], "cross_compile_flag": 71, "cross_compil": 71, "refit_module_weight": [71, 102], "compiled_modul": [71, 102], "new_weight_modul": [71, 102], "verify_output": [71, 102], "use_weight_map_cach": [71, 102], "in_plac": [71, 102], "compmil": 71, "coverag": [71, 96], "min_acc_module_s": 72, "is_aten": 72, "use_experimental_fx_rt": 72, "correctness_atol": 72, "correctness_rtol": 72, "minim": [72, 91, 96, 101], "submodul": [72, 88, 96, 108], "fx2trt": 72, "cpu": [72, 101, 109, 110, 111, 113], "has_batch_dim": 72, "dtyep": 72, "prop": 72, "min_input_shap": 72, "optimized_input_shap": 72, "max_input_shap": 72, "popul": 72, "225": [72, 114, 115, 117], "explicit_precis": 72, "logger_level": 72, "model_trt": [73, 93], "model_torchtrt": 73, "internal_error": 73, "dataloadercalibr": [74, 91], "preprocess": [74, 91, 114, 115, 117], "algo_typ": [74, 91], "calibrationalgo": [74, 91], "cachecalibr": [74, 91], "qualnam": [74, 76], "entropy_calibr": 74, "entropy_calibration_2": [74, 91], "legacy_calibr": 74, "minmax_calibr": 74, "set_multi_device_safe_mod": [75, 120], "_multidevicesafemodecontextmanag": 75, "impact": 75, "suppress": 75, "unsaf": 75, "trt_compiled_modul": 75, "torchtensorrtmodul": [75, 96], "encompass": [75, 77], "simpili": 75, "de": 75, "initi": [75, 76, 82, 102, 103, 105, 107, 108, 109, 110], "scriptmodul": [75, 76, 77, 89, 90, 121, 122], "overridden": [75, 76], "subclass": 75, "although": [75, 82], "recip": [75, 91], "afterward": 75, "former": 75, "care": 75, "hook": 75, "silent": 75, "get_extra_st": 75, "state_dict": [75, 76, 100], "set_extra_st": 75, "picklabl": 75, "pickl": [75, 96, 98], "load_state_dict": [75, 100, 112], "pythontorchtensorrtmodul": 75, "serialized_engin": [75, 77], "_set": [75, 103], "weight_name_map": 75, "trt_modul": [75, 120], "engine_str": 75, "my_modul": 75, "current_devic": 75, "disable_profil": 75, "enable_profil": 75, "iprofil": 75, "spent": 75, "get_layer_info": 75, "validate_input_shap": 75, "request": [76, 89, 114, 115, 117], "decid": 76, "deseri": [76, 77, 89, 96], "retrac": 76, "cudagraphstorchtensorrtmodul": 76, "strict": [76, 111, 120], "valueerror": [76, 95], "mutabletorchtensorrtmodul": [76, 100], "pytorch_model": 76, "regular": 76, "whenev": 76, "refit_gm": 76, "shape_mod": 76, "_shapemod": 76, "interv": 76, "notat": 76, "bound": 76, "torch_tensor": 76, "tracer": 76, "example_tensor": 76, "optimization_profile_field": 76, "classmethod": 76, "disable_memory_format_check": 76, "core_id": 76, "schedul": [76, 114, 115, 117], "use_default": 76, "try_to": 76, "anoth": [76, 82, 83, 88, 90, 102], "typeerror": 76, "unknown": 76, "succe": 76, "float_dtyp": 76, "failur": 76, "bf16": 76, "try_from": [76, 96], "complex128": 76, "16": [76, 86, 88, 89, 90, 105, 108], "brain": 76, "bfloat16": 76, "f64": 76, "f8": 76, "fp8": 76, "float8": 76, "i32": 76, "sign": [76, 114, 115, 117], "i64": 76, "u8": 76, "unsign": 76, "uint8": [76, 111], "trt_dla": 76, "torchtrt_dla": 76, "_from": 76, "torchtrt_dla_ec": 76, "torchtrt_safety_ec": 76, "saefti": 76, "trt_dla_ec": 76, "standalon": [76, 82, 111], "certifi": 76, "tf": 76, "torchtrt_linear": 76, "cdhw32": 76, "thirti": 76, "row": [76, 83], "spatial": 76, "31": [76, 89], "subscript": [76, 82], "chw16": 76, "sixteen": 76, "15": [76, 82, 86], "chw2": 76, "chw32": 76, "chw4": 76, "four": [76, 82, 83], "dhwc": 76, "equivi": 76, "channels_last_3d": 76, "dhwc8": 76, "eight": 76, "dla_hwc4": 76, "imag": [76, 91, 96, 100, 106, 112, 114, 115, 117], "roundup": 76, "elements": 76, "dla_linear": 76, "planar": 76, "hwc": 76, "channels_last": 76, "hwc16": 76, "hwc8": 76, "least": [76, 82, 83], "ishapelay": 77, "check_method_op_support": 77, "seriali": 77, "put_binding_nam": 77, "tensorrtcompilespec": [77, 92], "scriptclass": 77, "0x7f70888fd830": 77, "_jit_to_tensorrt": 77, "00": 78, "000": [78, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113], "total": [78, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113], "galleri": [78, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114], "mem": 78, "torch_compile_advanced_usag": [78, 103], "torch_compile_resnet_exampl": [78, 105], "torch_compile_stable_diffus": [78, 106], "torch_compile_transformers_exampl": [78, 107], "v0": [79, 114, 115, 117], "pytorch_sphinx_them": [80, 87], "conf": [80, 87], "html_theme_opt": 80, "canonical_url": 80, "analytics_id": 80, "logo_onli": 80, "display_vers": 80, "prev_next_buttons_loc": 80, "bottom": 80, "style_external_link": 80, "vcs_pageview_mod": 80, "collapse_navig": 80, "sticky_navig": [80, 84], "navigation_depth": 80, "includehidden": 80, "titles_onli": 80, "canon": 80, "rank": 80, "trail": 80, "slash": 80, "googl": 80, "analyt": 80, "isn": [80, 82, 93, 96], "shown": [80, 82, 89, 111, 119], "sidebar": [80, 86], "button": [80, 82], "icon": [80, 82], "extern": [80, 82, 99, 114], "display_github": 80, "display_gitlab": 80, "gitlab": 80, "bitbucket": 80, "bar": [80, 82], "www": [80, 82, 89, 91, 114, 115, 117], "sphinx": [80, 81, 82, 83, 87, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114], "toctre": 80, "lose": 80, "scroll": [80, 84], "unlimit": 80, "header": [80, 82, 83, 89, 114, 115, 117], "render": 80, "github_url": 80, "bitbucket_url": 80, "gitlab_url": 80, "left": [80, 82], "upon": [80, 103, 107], "rst": [80, 82], "visitor": 80, "revert": 80, "misbuild": 80, "properti": [80, 96], "stick": 80, "screen": 80, "vertic": [80, 82], "too": [80, 82, 83], "sticki": [80, 86], "nav": [80, 86], "At": [81, 94, 102], "django": 81, "payment": 81, "dotpai": 81, "dotpayprovid": 81, "seller_id": 81, "pin": 81, "lock": 81, "lang": 81, "pl": 81, "polish": 81, "gatewai": 81, "transfer": 81, "purchas": 81, "item": [81, 83, 112], "param": 81, "seller": 81, "consult": 81, "ui": 81, "languag": [81, 82, 83, 88, 93, 96, 99, 104, 109, 114, 115, 117], "data_item_1": 81, "emphasi": 82, "hyperlink": 82, "uri": 82, "web": 82, "anonym": 82, "label": [82, 91, 111, 112, 114, 115, 116, 117], "substitut": 82, "charact": 82, "exceedingli": 82, "ugli": 82, "problem": [82, 110], "problemat": 82, "ext": [82, 83], "autodoc": [82, 83], "demo": [82, 91, 111], "test_py_modul": [82, 86], "my": [82, 104, 109], "role": 82, "pep": 82, "287": 82, "rfc": 82, "2822": 82, "superscript": 82, "gui": 82, "taken": [82, 101], "height": [82, 111], "interfer": 82, "press": 82, "keyboard": 82, "mous": 82, "mmb": 82, "menuselect": 82, "seen": [82, 83], "whitespac": 82, "signific": [82, 96], "strang": 82, "hyphen": 82, "word": [82, 116], "adjust": 82, "width": [82, 111, 116], "browser": 82, "sentenc": [82, 113, 116], "suppli": [82, 102], "258": 82, "equat": 82, "x_": 82, "x_0": 82, "x_1": 82, "x_2": 82, "x_3": 82, "x_4": 82, "nabla": 82, "frac": 82, "theta": 82, "phi": 82, "restructuredtext": [82, 83], "parser": [82, 95, 112], "colon": 82, "indent": 82, "literal_block": 82, "spaces_and_linebreak": 82, "preserv": [82, 88, 91, 111], "markup_process": 82, "Or": 82, "great": [82, 89, 93, 96, 98, 122], "why": [82, 120], "didn": 82, "blank": 82, "align": 82, "permit": 82, "awai": 82, "eric": 82, "orchestra": 82, "leader": 82, "bee": 82, "philosoph": 82, "ipso": 82, "facto": 82, "But": [82, 89, 102, 113], "got": [82, 89], "vi": 82, "entiti": 82, "said": 82, "entir": [82, 108, 111, 122], "ancient": 82, "injuri": 82, "sing": 82, "elk": 82, "bracket": 82, "miss": [82, 89], "brontosaurus": 82, "thin": 82, "thicker": 82, "middl": 82, "That": [82, 89], "mine": 82, "belong": 82, "me": [82, 83], "ann": 82, "begun": 82, "past": 82, "pars": [82, 89], "someurl": 82, "dev0": 82, "70e2a38": 82, "caption": [82, 85], "pane": 82, "shell_command": 82, "echo": 82, "did": 82, "window_nam": 82, "session_nam": 82, "shorthand": 82, "some_funct": 82, "highlight": 82, "THE": 82, "heaven": 82, "hexagram": 82, "six": 82, "unbroken": 82, "primal": 82, "light": [82, 121], "spirit": 82, "weak": 82, "essenc": 82, "energi": 82, "unrestrict": 82, "conceiv": 82, "motion": 82, "regard": [82, 122], "basi": 82, "thu": 82, "persist": 82, "dual": 82, "sens": [82, 89], "univers": 82, "world": 82, "men": 82, "express": 82, "deiti": 82, "human": 82, "denot": [82, 96], "holi": 82, "man": [82, 83], "sage": 82, "ruler": 82, "who": 82, "awaken": 82, "utf": [82, 83], "sphinx_rtd_them": [82, 83], "docstr": [82, 83, 90], "dl": 82, "dt": 82, "tag": [82, 114, 115, 117], "tt": 82, "descnam": 82, "descclassnam": 82, "wrote": 82, "anyth": [82, 83, 111, 120], "programm": 82, "myclass": 82, "dothismethod": 82, "flush": 82, "meth": 82, "capit": 82, "flox": 82, "unreferenc": 82, "nonexist": 82, "extrem": 82, "stuff": 82, "mayb": 82, "bold": 82, "ital": 82, "heck": 82, "put": [82, 93, 116], "13": [82, 86], "backlink": 82, "knowledg": 82, "mind": 82, "ey": 82, "thought": 82, "medium": 82, "peopl": 82, "subsect": 82, "interpol": 82, "indirect": 82, "phrase": 82, "docutil": [82, 83], "sourceforg": [82, 83], "ref": 82, "clickabl": 82, "legend": 82, "revis": [82, 83, 100, 106], "revisit": 82, "enhanc": [82, 101, 111], "structuredtext": 82, "wooden": 82, "nickel": 82, "mad": 82, "scientist": 82, "bigger": 82, "bread": 82, "box": [82, 111, 118, 122], "wash": 82, "behind": 82, "ear": 82, "room": 82, "closet": 82, "bathroom": 82, "trash": 82, "sink": 82, "mother": 82, "g_": 82, "mu": 82, "nu": 82, "pi": 82, "t_": 82, "rho_": 82, "servic": 82, "thing1": 82, "thing2": 82, "thing3": 82, "prose": 82, "provok": 82, "mental": 82, "exert": 82, "reader": 82, "discret": 82, "strongli": [82, 113], "advis": 82, "subtitl": 82, "outsid": 82, "often": 82, "besid": 82, "border": [82, 111], "background": [82, 88], "ok": [82, 89], "transmit": 82, "disconnect": 82, "nonetheless": 82, "semant": 82, "blue": [82, 96], "white": [82, 111], "arab": 83, "roman": 83, "upper": 83, "iii": 83, "iv": 83, "classifi": [83, 88, 89, 112, 116], "paragraph": [83, 86], "z": [83, 93], "commonli": 83, "vm": 83, "david": 83, "goodger": 83, "address": [83, 96, 100], "123": 83, "street": 83, "canada": 83, "a1b": 83, "2c3": 83, "contact": 83, "myself": 83, "organ": 83, "humankind": 83, "2012": 83, "03": 83, "19": [83, 86], "53": 83, "0000": 83, "tue": 83, "jan": 83, "progress": 83, "7302": 83, "wish": 83, "redistribut": 83, "reattribut": 83, "sell": 83, "bui": 83, "rent": 83, "leas": 83, "improv": [83, 101, 108, 111, 120], "quot": 83, "excerpt": 83, "collat": 83, "fold": 83, "stapl": 83, "mutil": 83, "anyon": 83, "heart": 83, "bibliograph": 83, "markup": [83, 86], "literal": 83, "yahoo": 83, "oh": 83, "liter": 83, "heh": 83, "child": 83, "beat": 83, "text": [83, 85, 104, 109, 110, 116], "hehe": 83, "kept": 83, "sai": [83, 116], "cackl": 83, "night": 83, "lone": 83, "guangzhou": 83, "destini": 83, "hope": 83, "dream": 83, "forth": 83, "fifth": 83, "sixth": 83, "lorem": [83, 85], "ipsum": [83, 85], "dolor": [83, 85], "sit": [83, 85], "amet": [83, 85], "consectetur": [83, 85], "adipisc": [83, 85], "elit": [83, 85], "donec": [83, 85], "porttitor": [83, 85], "odio": [83, 85], "posuer": [83, 85], "vita": [83, 85], "ornar": [83, 85], "libero": [83, 85], "matti": 83, "loborti": [83, 85], "justo": [83, 85], "vestibulum": [83, 85], "nibh": [83, 85], "aliquet": [83, 85], "feugiat": [83, 85], "sagitti": [83, 85], "nequ": [83, 85], "qui": [83, 85], "eleifend": 83, "dui": [83, 85], "rutrum": [83, 85], "lectu": [83, 85], "suscipit": [83, 85], "letter": [83, 116], "column": 83, "cell": 83, "span": 83, "nam": [83, 85], "mauri": [83, 85], "arcu": [83, 85], "stub": 83, "behav": 84, "area": 84, "interdum": 85, "nec": 85, "finibu": 85, "dictum": 85, "velit": 85, "ut": 85, "eu": 85, "efficitur": 85, "aliquam": 85, "erat": 85, "diam": 85, "gravida": 85, "imperdiet": 85, "tellu": 85, "nisl": 85, "praesent": 85, "eget": 85, "elementum": 85, "rhoncu": 85, "tincidunt": 85, "suspendiss": 85, "volutpat": 85, "scelerisqu": 85, "tristiqu": 85, "aenean": 85, "condimentum": 85, "risu": 85, "accumsan": 85, "laoreet": 85, "maximu": 85, "sapien": 85, "ligula": 85, "fringilla": 85, "commodo": 85, "proin": 85, "et": 85, "pharetra": 85, "etiam": 85, "turpi": 85, "ant": 85, "luctu": 85, "vel": 85, "malesuada": 85, "dignissim": 85, "mi": 85, "nunc": 85, "augu": 85, "sem": 85, "cursu": 85, "nulla": 85, "pellentesqu": 85, "habit": 85, "morbi": 85, "senectu": 85, "netu": 85, "fame": 85, "ac": 85, "egesta": 85, "placerat": 85, "tortor": 85, "iaculi": 85, "venenati": 85, "cra": 85, "puru": 85, "ero": 85, "vehicula": 85, "fusc": 85, "auctor": 85, "phasellu": 85, "est": 85, "viverra": 85, "conval": 85, "faucibu": 85, "vulput": 85, "feli": 85, "sodal": 85, "maecena": 85, "congu": 85, "semper": 85, "enim": 85, "blandit": 85, "sollicitudin": 85, "urna": 85, "orci": 85, "lacu": 85, "quisqu": 85, "facilisi": 85, "hendrerit": 85, "curabitur": 85, "variu": 85, "bibendum": 85, "massa": 85, "magna": 85, "tempu": 85, "metu": 85, "nisi": 85, "pretium": 85, "leo": 85, "euismod": 85, "ultric": 85, "dapibu": 85, "lacinia": 85, "vivamu": 85, "molesti": 85, "hac": 85, "habitass": 85, "platea": 85, "dictumst": 85, "git": 86, "content": [86, 91, 114, 115, 117], "changelog": 86, "math": 86, "14": [86, 97, 107, 114, 115, 117], "17": 86, "18": [86, 89, 100, 111], "submenu": 86, "symlink": 87, "subtre": 87, "_theme": 87, "html_theme": 87, "html_theme_path": 87, "optimiz": 88, "tutori": [88, 91, 93, 94, 96, 98, 100, 102, 115, 117], "beginn": 88, "intro_to_torchscript_tutori": 88, "briefli": 88, "lenet": [88, 89], "lenetfeatextractor": 88, "conv1": [88, 89], "conv2d": [88, 96, 112], "conv2": [88, 89], "lenetclassifi": 88, "fc1": [88, 89], "120": [88, 89], "fc2": [88, 89], "84": [88, 89], "fc3": [88, 89], "feat": [88, 89, 111], "obvious": 88, "pathwai": 88, "input_data": [88, 90], "traced_model": 88, "pick": [88, 119], "script_model": [88, 92], "perspect": 88, "___torch_mangle_10": 88, "129": 88, "___torch_mangle_9": 88, "119": 88, "___torch_mangle_5": 88, "137": 88, "callmethod": 88, "138": 88, "38": 88, "39": 88, "torch_script_modul": [88, 89], "in_tensor": 88, "fly": 88, "lenet_script": [88, 89], "haven": 89, "acquir": 89, "dyanmo": 89, "almost": [89, 122], "trt_lenet_script": 89, "apr": 89, "56": 89, "04": 89, "credit": 89, "stop": 89, "argc": 89, "argv": 89, "cerr": 89, "cout": 89, "even": [89, 100, 108], "cppdoc": 89, "pretti": 89, "fashion": [89, 116], "enable_precis": 89, "And": 89, "convertgraphtotrtengin": 89, "engine_converted_from_jit": 89, "close": [89, 94, 111], "saw": 89, "576": 89, "346": 89, "539": 89, "0464": 89, "0383": 89, "0678": 89, "0932": 89, "1045": 89, "0805": 89, "0435": 89, "0818": 89, "0208": 89, "0358": 89, "cudafloattyp": 89, "0530": 89, "1691": 89, "2802": 89, "1502": 89, "1056": 89, "1549": 89, "input0": [89, 90], "1063": 89, "input1": [89, 90], "input2": 89, "28": 89, "29": 89, "33": 89, "35": 89, "36": 89, "37": 89, "compilegraph": [89, 91], "transform": [89, 91, 93, 97, 99, 101, 102, 104, 107, 109, 110, 111, 112, 113, 114, 115, 117, 121], "laid": 89, "translat": [89, 102], "aren": 89, "techniqu": [89, 91, 110, 120], "checkmethodoperatorsupport": 89, "modular": 89, "ship": [89, 120], "exhaust": 89, "109": 89, "addlay": 89, "yourself": 89, "question": [89, 94], "outself": 89, "flatten_convert": 89, "unwraptoint": 89, "in_shap": 89, "tovec": 89, "out_shap": 89, "shuffl": [89, 91, 112], "addshuffl": 89, "setreshapedimens": 89, "todim": 89, "extens": [89, 122], "ctype": 89, "cdll": 89, "contributor": 89, "upstream": 89, "pr": 89, "usecas": 90, "sole": [90, 91, 122], "individu": 90, "accuraci": [91, 111, 116], "loss": [91, 116], "infrastructur": [91, 114, 115, 117], "streamlin": [91, 93], "expos": [91, 96], "cpp_frontend": 91, "loading_data_recip": 91, "cifar10": [91, 112], "cstddef": 91, "ktrain": 91, "ktest": 91, "un": 91, "cs": 91, "toronto": 91, "edu": 91, "kriz": 91, "cifar": 91, "is_train": 91, "trim": 91, "use_subset": 91, "new_siz": 91, "mode_": 91, "images_": 91, "targets_": 91, "calibration_dataset": 91, "data_dir": 91, "320": 91, "4914": [91, 112], "4822": [91, 112], "4465": [91, 112], "2023": [91, 112], "1994": [91, 112], "2010": [91, 112], "dataloaderopt": 91, "worker": 91, "virtual": 91, "input_shap": [91, 123], "compile_spec": [91, 95, 105, 123], "kf16": [91, 123], "ki8": 91, "vgg16": [91, 112], "testing_dataset": [91, 112], "totensor": [91, 112, 114, 115, 117], "testing_dataload": [91, 112], "num_work": [91, 112], "vgg": [91, 112], "test_ptq_dataloader_calibr": 91, "test_ptq_trt_calibr": 91, "krizhevski": 91, "hinton": 91, "2009": 91, "tini": 91, "simonyan": 91, "zisserman": 91, "2014": 91, "recognit": [91, 116], "arxiv": 91, "preprint": 91, "1409": 91, "1556": 91, "_jit_to_backend": 92, "mobilenet_v2": 92, "pretrain": [92, 98, 100, 104, 105, 108, 114, 115, 116, 117], "cost": [93, 96, 98, 102, 120], "perhap": [93, 96], "overhead": [93, 96, 101, 108, 113, 120], "involv": [93, 101, 102, 108], "greatli": 93, "perviou": 93, "elementwis": [93, 94], "launch": [93, 96, 108, 114, 115, 117], "tensorrt_bind": 93, "trtp": 93, "tl": [93, 96], "elementwise_mul_kernel": 93, "block_siz": [93, 96], "thread": [93, 120], "pid": [93, 96], "program_id": [93, 96], "block_start": 93, "offset": 93, "x_val": 93, "y_val": 93, "wise": 93, "z_val": 93, "custom_op": [93, 96], "torchtrt_ex": [93, 96], "elementwise_mul": 93, "mutates_arg": [93, 96], "assert": [93, 96, 100, 102], "is_cuda": 93, "empty_lik": 93, "grid": 93, "cours": 93, "register_fak": [93, 96], "creation": 93, "less": 93, "boilerpl": [93, 94], "tensordesc": 93, "prior": [93, 94, 98, 118, 120], "x_t": 93, "as_tensor": [93, 96, 111], "y_t": 93, "z_t": 93, "generate_plugin_convert": 93, "supports_dynamic_shap": [93, 94], "my_model": [93, 96], "allclos": [93, 94, 100, 102], "ran": 93, "minut": [93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113], "auto_generate_convert": 93, "jupyt": [93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114], "ipynb": [93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113], "gelu": 94, "sy": 94, "approxim": 94, "suppos": 94, "my_mod": 94, "ex_input": [94, 96], "baselin": 94, "my_standard_gelu": 94, "supersed": 94, "converterprior": 94, "vers": 94, "distinct": 94, "prepend": 94, "candid": 94, "primit": 94, "compiler_ir": 94, "focu": [94, 100], "interoper": 94, "aten_ops_gelu": 94, "sourceir": 94, "cheap": 94, "unqiu": 94, "op_count": 94, "get_op_count": 94, "nonloc": 94, "source_ir": 94, "lhs_val": 94, "rhs_val": 94, "x_7": 94, "x_8": 94, "79788456080000003": 94, "x_9": 94, "044714999999999998": 94, "x_10": 94, "x_11": 94, "x_12": 94, "x_13": 94, "x_14": 94, "x_15": 94, "my_custom_gelu": 94, "my_mod_erf": 94, "my_gelu_erf": 94, "notic": [94, 101], "converter_overload": 94, "geforcertx": 95, "4080": 95, "3080": 95, "cross_runtime_compilation_for_window": 95, "trt_resnet": 95, "argpars": [95, 112], "argumentpars": [95, 112], "comil": 95, "add_argu": [95, 112], "parse_arg": [95, 112], "manual_se": [95, 97, 98, 100, 102], "resnet18": [95, 98, 100, 102, 105, 108], "amd64": 95, "loaded_model": 95, "load_cross_compiled_exported_program": 95, "trt_output": 95, "cross_compile_for_window": 95, "sake": 96, "circular": 96, "red": [96, 111], "green": [96, 111], "twice": 96, "written": 96, "openai": 96, "formal": 96, "circ_pad_kernel": 96, "all_pads_0": 96, "all_pads_2": 96, "all_pads_4": 96, "all_pads_6": 96, "orig_dims_0": 96, "orig_dims_1": 96, "orig_dims_2": 96, "orig_dims_3": 96, "y_shape_1": 96, "y_shape_2": 96, "y_shape_3": 96, "x_len": 96, "y_len": 96, "mask_i": 96, "i3": 96, "i2": 96, "i1": 96, "i0": 96, "j0": 96, "j1": 96, "j2": 96, "j3": 96, "load_idx": 96, "mask_x": 96, "triton_circular_pad": 96, "out_dim": 96, "tolist": 96, "all_pad": 96, "zero": 96, "orig_dim": 96, "blocksiz": 96, "256": [96, 111, 112, 113, 114, 115, 117], "numblock": 96, "tracabl": 96, "prerequisit": 96, "fake": 96, "real": 96, "faketensor": 96, "autograd": 96, "beyond": 96, "register_autograd": 96, "padded_x": 96, "2604": 96, "4232": 96, "3041": 96, "0833": 96, "2461": 96, "1270": 96, "2450": 96, "4079": 96, "2887": 96, "2828": 96, "0373": 96, "0332": 96, "3143": 96, "6344": 96, "5638": 96, "1867": 96, "5068": 96, "4363": 96, "7937": 96, "3488": 96, "1350": 96, "7966": 96, "3517": 96, "1379": 96, "5537": 96, "1088": 96, "8950": 96, "0550": 96, "6163": 96, "0109": 96, "5245": 96, "9632": 96, "5686": 96, "3775": 96, "8162": 96, "4216": 96, "4311": 96, "1649": 96, "2091": 96, "3668": 96, "1006": 96, "1447": 96, "0352": 96, "7689": 96, "8131": 96, "_run_on_gpu_0": 96, "_run_on_acc_1": 96, "dry": 96, "50": [96, 116], "count": 96, "__": 96, "aggreg": 96, "stat": 96, "latenc": [96, 111, 113, 120], "abstractli": 96, "pkl": [96, 100], "cupi": 96, "gap": 96, "prealloc": 96, "circularpaddingplugin": 96, "ipluginv2dynamicext": 96, "field_collect": 96, "pluginfieldcollect": 96, "x_shape": 96, "num_output": 96, "plugin_namespac": 96, "plugin_typ": 96, "plugin_vers": 96, "get_output_datatyp": 96, "input_typ": 96, "get_output_dimens": 96, "output_index": 96, "dimsexpr": 96, "exprbuild": 96, "iexprbuild": 96, "output_dim": 96, "dimensionoper": 96, "configure_plugin": 96, "inp": 96, "dynamicplugintensordesc": 96, "x_dim": 96, "desc": 96, "supports_format_combin": 96, "po": 96, "in_out": 96, "plugintensordesc": 96, "num_input": 96, "enqueu": 96, "input_desc": 96, "output_desc": 96, "in_dtyp": 96, "a_mem": 96, "unownedmemori": 96, "items": 96, "c_mem": 96, "a_ptr": 96, "memorypoint": 96, "c_ptr": 96, "a_d": 96, "memptr": 96, "c_d": 96, "a_t": 96, "c_t": 96, "cloned_plugin": 96, "__dict__": 96, "circularpaddingplugincr": 96, "iplugincr": 96, "field_nam": 96, "pluginfield": 96, "pluginfieldtyp": 96, "create_plugin": 96, "pluginfieldcollection_": 96, "deserialize_plugin": 96, "pads_dict": 96, "creator": 96, "trt_plugin_registri": 96, "get_plugin_registri": 96, "register_cr": 96, "untyp": 96, "get_trt_tensor": 96, "set_layer_nam": 96, "recal": 96, "intlist": 96, "circular_padding_convert": 96, "retriev": 96, "elsewher": 96, "plugin_registri": 96, "plugin_cr": 96, "get_plugin_cr": 96, "field_config": 96, "eventu": 96, "freez": 96, "_input": 96, "add_plugin_v2": 96, "circular_padding_plugin": 96, "_run_on_acc_0": 96, "grad_fn": 96, "subbackward0": 96, "custom_kernel_plugin": 96, "engine_caching_exampl": [97, 98], "remove_timing_cach": [97, 98], "bertmodel": [97, 101, 107], "random": [97, 98, 100, 102, 111, 113], "seed": [97, 98, 100, 102], "from_pretrain": [97, 100, 101, 104, 106, 107, 109, 110, 111, 113], "uncas": [97, 101, 107, 116], "return_dict": 97, "randint": [97, 101, 107, 113], "compile_bert": 97, "enable_tim": [97, 98], "1st": [97, 98], "measur": [97, 98, 113], "2nd": [97, 98], "3rd": [97, 98], "slower": [97, 98], "messur": [97, 98], "compilation_kwarg": [97, 107], "torch_trt_bert_engine_cach": 97, "30": [97, 98, 100, 102, 105, 107, 111, 119], "synchron": [97, 98, 101, 113], "elapsed_tim": [97, 98], "millisecond": 97, "__name__": [97, 103, 107], "__main__": [97, 103, 107], "engine_caching_bert_exampl": 97, "paid": 98, "upfront": 98, "invalid": 98, "repeatedli": 98, "mitig": [98, 101], "explor": 98, "torch_trt": [98, 100, 102], "_default": 98, "_engine_cach": 98, "flexibl": [98, 122], "histor": 98, "barrier": 98, "reconstruct": 98, "ti": 98, "hash": 98, "magnitud": 98, "torch_compil": [98, 103, 105, 107, 108, 118, 122], "compiled_model": 98, "ms": [98, 101, 113], "dynamo_compil": 98, "example_input": 98, "200": 98, "dynamic_shap": [98, 118], "remot": 98, "systen": 98, "agnost": 98, "implent": 98, "ramenginecach": 98, "held": 98, "engine_cach": 98, "torch_compile_my_cach": 98, "_torch_compile_gpt2": [99, 114], "_torch_export_gpt2": [99, 114], "_torch_export_llama2": [99, 114], "_torch_export_sam2": [99, 114], "sphx_glr_tutorials__rendered_examples_dynamo_cross_runtime_compilation_for_window": [99, 114], "straightforward": 100, "especi": [100, 101], "hug": [100, 104, 109, 110], "face": [100, 104, 109, 110], "difficult": 100, "ever": [100, 104], "walk": [100, 102, 104, 109], "lora": [100, 102], "use_python": 100, "mutable_modul": 100, "model2": [100, 102], "expected_output": [100, 102], "refitted_output": [100, 102], "reload": [100, 122], "checkpoint": [100, 112], "civitai": 100, "12597": 100, "moxin": 100, "diffusionpipelin": [100, 106], "no_grad": [100, 101, 104, 109, 110, 112, 113], "model_id": [100, 106], "runwayml": 100, "hous": 100, "forest": 100, "shuimobysim": 100, "wuchangshuo": 100, "qualiti": 100, "worst": 100, "lowr": 100, "cloudi": 100, "watermark": 100, "pipe": [100, 106], "torch_dtyp": [100, 106], "unet": [100, 106], "negative_prompt": 100, "num_inference_step": 100, "without_lora_mut": 100, "jpg": [100, 111, 114, 115, 117], "procedur": 100, "load_lora_weight": 100, "stablediffusionapi": 100, "load_lora_embed": 100, "weight_nam": 100, "safetensor": 100, "adapter_nam": 100, "lora1": 100, "set_adapt": 100, "adapter_weight": 100, "fuse_lora": 100, "unload_lora_weight": 100, "with_lora_mut": 100, "mutable_torchtrt_module_exampl": 100, "act": 101, "concurr": [101, 114, 115, 117], "overlap": 101, "particularli": 101, "cycl": 101, "overal": [101, 116], "workload": 101, "enough": 101, "overshadow": 101, "cumul": 101, "priorit": 101, "comprehens": 101, "infrequ": 101, "timeit": [101, 113], "test_module_perf": 101, "warm": [101, 108, 113], "accur": 101, "start_tim": [101, 113], "default_tim": [101, 113], "end_tim": [101, 113], "time_m": 101, "median": 101, "metric": 101, "128": [101, 111, 112, 113], "enable_pre_allocated_output": 101, "out_trt": [101, 108], "pre_allocated_output_ctx": 101, "set_pre_allocated_output": 101, "time_opt": 101, "time_norm": 101, "time_opt_m": 101, "1000": [101, 112, 113, 114, 115, 117], "time_normal_m": 101, "3f": [101, 111], "pre_allocated_output_exampl": 101, "expens": 102, "occasion": [102, 103, 107], "adapt": 102, "infeas": 102, "focus": 102, "mostli": 102, "recogn": 102, "behalf": 102, "init": [102, 112], "sett": 102, "randomli": 102, "exp_program2": 102, "compiled_trt_ep": 102, "new_trt_gm": 102, "accomplish": 102, "gaurente": 102, "attempt": [102, 112, 118], "rebuild": 102, "heurist": 102, "refit_engine_exampl": 102, "x_out": 103, "y_out": 103, "x_y_out": 103, "invoc": 103, "sample_inputs_half": 103, "model_half": 103, "backend_kwarg": 103, "optimized_model_custom": 103, "exit": [103, 107], "2052": [103, 107], "compile_engine_and_inf": [103, 107], "art": [104, 111], "causal": 104, "unidirect": 104, "corpu": [104, 116], "huggingfac": [104, 109, 110, 116], "automodelforcausallm": [104, 109, 110, 113], "autotoken": [104, 109, 110], "success": 104, "max_length": 104, "token": [104, 109, 110, 116], "kv_cach": [104, 109, 110], "pad_token_id": [104, 109], "eos_token_id": [104, 109, 110], "attn_implement": [104, 109, 110, 113], "eager": [104, 109, 110, 113], "enjoi": [104, 109], "cute": [104, 109], "dog": [104, 109], "model_input": [104, 109, 110], "return_tensor": [104, 109, 110], "input_id": [104, 109, 110], "regress": [104, 109, 110], "pyt_gen_token": [104, 109, 110], "mark_dynam": [104, 105, 118], "1023": 104, "trt_gen_token": [104, 109, 110], "skip_special_token": [104, 109, 110], "torch_compile_gpt2": 104, "new_input": [105, 107], "new_output": [105, 107], "new_batch_size_input": 105, "new_batch_size_output": 105, "inputs_bs8": 105, "outputs_bs8": 105, "No": [105, 118], "inputs_bs12": 105, "outputs_bs12": 105, "compvi": 106, "majest": 106, "castl": 106, "cloud": 106, "majestic_castl": 106, "png": [106, 111], "enable_cudagraph": [108, 120], "cudagraphs_modul": 108, "set_cudagraphs_mod": [108, 120], "inputs_2": 108, "inputs_3": 108, "out_trt_2": 108, "out_trt_3": 108, "diminish": 108, "encapsul": 108, "wrapped_modul": 108, "captur": 108, "replai": 108, "samplemodel": 108, "intention": 108, "Of": 108, "manner": 108, "opt_with_graph_break": 108, "torch_export_cudagraph": 108, "export_llm": [109, 110, 113], "max_token": [109, 110, 113], "gpt2_ep": 109, "max_seq_len": [109, 110, 113], "parallel": 109, "paradigm": 109, "torch_export_gpt2": 109, "llama_path": [110, 113], "llama": [110, 113], "7b": [110, 113], "chat": [110, 113], "hf": [110, 113], "llama2_ep": [110, 113], "batch_decod": 110, "clean_up_tokenization_spac": 110, "solv": [110, 111, 114, 115, 117], "smaller": [110, 116], "subproblem": 110, "torch_export_llama2": 110, "foundat": 111, "promptabl": 111, "video": 111, "fork": 111, "condition": 111, "concaten": 111, "layernorm": 111, "reli": 111, "stabil": 111, "matplotlib": 111, "pyplot": 111, "plt": 111, "panda": 111, "pd": 111, "pil": [111, 114, 115, 117], "sam2_image_predictor": 111, "sam2imagepredictor": 111, "sam_compon": 111, "sam2fullmodel": 111, "agg": 111, "facebook": 111, "hiera": 111, "set_imag": 111, "predict": 111, "predictor": 111, "image_encod": 111, "forward_imag": 111, "_prepare_backbone_featur": 111, "directly_add_no_mem_emb": 111, "no_mem_emb": 111, "_featur": 111, "prompt_encod": 111, "sam_prompt_encod": 111, "mask_decod": 111, "sam_mask_decod": 111, "_bb_feat_siz": 111, "point_coord": 111, "point_label": 111, "backbone_out": 111, "vision_feat": 111, "feat_siz": 111, "image_emb": 111, "high_res_feat": 111, "high_res_featur": 111, "feat_level": 111, "sparse_embed": 111, "dense_embed": 111, "low_res_mask": 111, "iou_predict": 111, "image_embed": 111, "image_p": 111, "get_dense_p": 111, "sparse_prompt_embed": 111, "dense_prompt_embed": 111, "multimask_output": 111, "repeat_imag": 111, "sam_model": 111, "input_imag": 111, "truck": 111, "rgb": 111, "sam2transform": 111, "facebookresearch": 111, "preprocess_input": 111, "orig_hw": 111, "_transform": 111, "500": 111, "375": 111, "unnorm_coord": 111, "transform_coord": 111, "postprocess": 111, "plot": 111, "confid": 111, "score": 111, "postprocess_mask": 111, "resolut": [111, 116], "sorted_indic": 111, "argsort": 111, "show_mask": 111, "ax": 111, "random_color": 111, "255": 111, "144": 111, "astyp": 111, "mask_imag": 111, "cv2": 111, "contour": 111, "findcontour": 111, "retr_extern": 111, "chain_approx_non": 111, "smooth": 111, "approxpolydp": 111, "epsilon": 111, "drawcontour": 111, "thick": 111, "imshow": 111, "show_point": 111, "coord": 111, "marker_s": 111, "pos_point": 111, "neg_point": 111, "marker": 111, "edgecolor": 111, "linewidth": 111, "visualize_mask": 111, "title_prefix": 111, "overlaid": 111, "figsiz": 111, "gca": 111, "titl": 111, "fontsiz": 111, "savefig": 111, "_output_mask_": 111, "snippet": 111, "torchtrt_input": 111, "unnormalized_coordin": 111, "foreground": 111, "trt_out": 111, "trt_mask": 111, "trt_score": 111, "sam": 111, "torch_export_sam2": 111, "modelopt": 112, "mtq": 112, "export_torch_mod": 112, "layer_spec": 112, "num_class": 112, "init_weight": 112, "in_channel": 112, "pool": [112, 123], "maxpool2d": 112, "batchnorm2d": 112, "sequenti": 112, "avgpool": 112, "adaptiveavgpool2d": 112, "4096": 112, "dropout": 112, "_initialize_weight": 112, "kaiming_normal_": 112, "fan_out": 112, "nonlinear": 112, "constant_": 112, "elif": 112, "normal_": 112, "vgg16_cfg": 112, "ckpt": 112, "model_state_dict": 112, "device_count": 112, "ordereddict": 112, "new_state_dict": 112, "forget": 112, "training_dataset": 112, "randomcrop": 112, "randomhorizontalflip": 112, "training_dataload": 112, "drop_last": 112, "crit": 112, "crossentropyloss": 112, "calibrate_loop": 112, "pred": 112, "5f": 112, "acc": 112, "2f": 112, "quantize_typ": 112, "quant_cfg": 112, "int8_default_cfg": 112, "fp8_default_cfg": 112, "forward_loop": 112, "qdq": 112, "incomplet": 112, "functionaltensor": 112, "functionaltensormod": 112, "_trace": 112, "_export": 112, "float8_e4m3fn": 112, "class_prob": 112, "class_pr": 112, "test_prob": 112, "test_pr": 112, "test_loss": 112, "test_acc": 112, "vgg16_ptq": 112, "overcom": 113, "throughput": 113, "sometim": [113, 118], "outweigh": 113, "slowdown": 113, "hardwar": [113, 123], "experi": 113, "balanc": 113, "time_gener": 113, "output_seq_length": 113, "seq_len": [113, 118], "llm": 113, "input_seq": 113, "inputs_copi": 113, "decod": 113, "logit": 113, "next_token_logit": 113, "next_token": 113, "time_mean_m": 113, "isl": 113, "osl": 113, "solut": 113, "insight": 113, "weight_streaming_ctx": 113, "weight_stream": 113, "mean_lat": 113, "percentag": 113, "weight_budget_pct": 113, "device_budget": 113, "total_device_budget": 113, "permiss": 113, "equal": 113, "proportion": 113, "streamabl": 113, "streamable_budget": 113, "requested_budget": 113, "get_automatic_weight_streaming_budget": 113, "weight_streaming_exampl": 113, "hand": [114, 115, 117], "consider": [114, 115, 117], "grpc": [114, 115, 117], "aforement": [114, 115, 117], "familiar": [114, 115, 117], "resnet50": [114, 115, 117], "torchhub": [114, 115, 117], "docker": [114, 115, 117], "login": [114, 115, 117], "xx": [114, 115], "yy": [114, 115, 117], "mm": [114, 115, 117], "publish": [114, 115, 117], "pwd": [114, 115, 117], "scratch_spac": [114, 115, 117], "nvcr": [114, 115, 117], "py3": [114, 115, 117], "hub": [114, 115, 117], "_validate_not_a_forked_repo": [114, 115, 117], "ts_trt_model": [114, 115, 117], "triton_exampl": [114, 115, 117], "model_repositori": [114, 115, 117], "rm": [114, 115, 117], "highli": [114, 115, 116, 117], "suggest": [114, 115, 117], "simplest": [114, 115, 117], "pbtxt": [114, 115, 117], "data_typ": [114, 115, 117], "type_fp32": [114, 115, 117], "exact": [114, 115, 117], "encourag": [114, 115, 117], "proce": [114, 115, 117], "8000": [114, 115, 117], "8001": [114, 115, 117], "8002": [114, 115, 117], "tritonserv": [114, 115, 117], "spin": [114, 115, 117], "proceed": [114, 115, 117], "flesh": [114, 115, 117], "img1": [114, 115, 117], "hakaimagazin": [114, 115, 117], "wp": [114, 115, 117], "gulf": [114, 115, 117], "bird": [114, 115, 117], "attrdict": [114, 115, 117], "pyindex": [114, 115, 117], "tritoncli": [114, 115, 117], "jump": [114, 115, 117], "firstli": [114, 115, 117], "resiz": [114, 115, 117], "httpclient": [114, 115, 117], "triton_to_np_dtyp": [114, 115, 117], "rn50_preprocess": [114, 115, 117], "img_path": [114, 115, 117], "img": [114, 115, 117], "centercrop": [114, 115, 117], "485": [114, 115, 117], "456": [114, 115, 117], "406": [114, 115, 117], "229": [114, 115, 117], "transformed_img": [114, 115, 117], "inferenceservercli": [114, 115, 117], "localhost": [114, 115, 117], "secondli": [114, 115, 117], "obtain": [114, 115, 116, 117, 121], "inferinput": [114, 115, 117], "set_data_from_numpi": [114, 115, 117], "binary_data": [114, 115, 117], "inferrequestedoutput": [114, 115, 117], "class_count": [114, 115, 117], "lastli": [114, 115, 117], "send": [114, 115, 117], "model_nam": [114, 115, 117], "inference_output": [114, 115, 117], "as_numpi": [114, 115, 117], "468750": [114, 115, 117], "90": [114, 115, 117], "523438": [114, 115, 117], "92": [114, 115, 117], "664062": [114, 115, 117], "429688": [114, 115, 117], "136": [114, 115, 117], "234375": [114, 115, 117], "confidence_scor": [114, 115, 117], "classification_index": [114, 115, 117], "_rendered_examples_python": 114, "_rendered_examples_jupyt": 114, "acoust": 116, "speech": 116, "quartznet": 116, "contextnet": 116, "subword": 116, "piec": 116, "excit": 116, "se": 116, "audio": 116, "transcrib": 116, "speedup": 116, "feedforward": 116, "cnn": 116, "uniformli": 116, "compound": 116, "coeffici": 116, "b0": 116, "english": 116, "supervis": 116, "walkthrough": 116, "adopt": 116, "mobilenetv2": 116, "classif": 116, "imagenet": 116, "imagenett": 116, "qat": 116, "simul": 116, "eagerli": 118, "swap": 118, "exactli": 118, "_tracer": 118, "queri": 118, "attn_weight": 118, "compiler_dynamic_shap": 118, "inputs_bs2": 118, "mymodul": 119, "linear1": 119, "linear2": 119, "linear3": 119, "40": 119, "__myl_mulsum_myl0_0": 119, "layertyp": 119, "kgen": 119, "__mye116_dconst": 119, "__myln_k_arg__bb1_2": 119, "tacticnam": 119, "__myl_mulsum_0xfa6c1858aea1b13b03f90165d7149ec6": 119, "streamid": 119, "__myl_addresmulsum_myl0_1": 119, "__mye131_dconst": 119, "addmm_constant_0": 119, "addmm_add_broadcast_to_same_shape_lhs_broadcast_constantfloat": 119, "__myln_k_arg__bb1_3": 119, "__myl_addresmulsum_0xb3915d7ebfe48be45b6d49083479e12f": 119, "__myl_addresmulsumadd_myl0_2": 119, "__mye146_dconst": 119, "addmm_2_constant_0": 119, "addmm_2_add_broadcast_to_same_shape_lhs_broadcast_constantfloat": 119, "addmm_1_constant_0": 119, "addmm_1_add_broadcast_to_same_shape_lhs_broadcast_constantfloat": 119, "__myl_addresmulsumadd_0xcdd0085ad25f5f45ac5fafb72acbffd6": 119, "__myl_mulsumaddcas_myl0_0": 119, "__mye112_dconst": 119, "__myl_mulsumaddcas_0xacf8f5dd9be2f3e7bb09cdddeac6c936": 119, "__myl_resmulsumaddcas_myl0_1": 119, "__mye127_dconst": 119, "addmm_1_add_broadcast_to_same_shape_lhs_broadcast_constanthalf": 119, "__myl_resmulsumaddcas_0x5a3b318b5a1c97b7d5110c0291481337": 119, "__myl_resmulsumadd_myl0_2": 119, "__mye142_dconst": 119, "__myl_resmulsumadd_0x3fad91127c640fd6db771aa9cde67db0": 119, "libtorchtrt_runtim": 120, "dl_open": 120, "ld_preload": 120, "load_librari": 120, "wl": 120, "ltorchtrt": 120, "torchtrt_runtime_exampl": 120, "libtorchtrt_plugin": 120, "neglig": 120, "alert": 120, "switch": 120, "mismatch": 120, "crash": 120, "sacrif": 120, "incur": 120, "intens": 120, "trt_ep": 121, "stai": 121, "trt_t": 121, "ergonom": 122, "deleg": 122, "believ": 122, "amen": 122, "artifact": 122, "pack": 122, "year": 122, "superset": 122, "codebas": 122, "immedi": 122, "traceabl": 122, "scriptabl": 122, "neural": 123, "deconvolut": 123, "scripted_model": 123}, "objects": {"": [[5, 0, 1, "c.STR", "STR"], [9, 0, 1, "c.TORCHTRT_API", "TORCHTRT_API"], [11, 0, 1, "c.TORCHTRT_HIDDEN", "TORCHTRT_HIDDEN"], [7, 0, 1, "c.TORCH_TENSORRT_MAJOR_VERSION", "TORCH_TENSORRT_MAJOR_VERSION"], [8, 0, 1, "c.TORCH_TENSORRT_MINOR_VERSION", "TORCH_TENSORRT_MINOR_VERSION"], [6, 0, 1, "c.TORCH_TENSORRT_PATCH_VERSION", "TORCH_TENSORRT_PATCH_VERSION"], [12, 0, 1, "c.TORCH_TENSORRT_VERSION", "TORCH_TENSORRT_VERSION"], [10, 0, 1, "c.XSTR", "XSTR"], [0, 1, 1, "_CPPv4N14torch_tensorrt8DataTypeE", "torch_tensorrt::DataType"], [0, 2, 1, "_CPPv4N14torch_tensorrt8DataType8DataTypeE5Value", "torch_tensorrt::DataType::DataType"], [0, 2, 1, "_CPPv4N14torch_tensorrt8DataType8DataTypeEN3c1010ScalarTypeE", "torch_tensorrt::DataType::DataType"], [0, 2, 1, "_CPPv4N14torch_tensorrt8DataType8DataTypeEv", "torch_tensorrt::DataType::DataType"], [0, 3, 1, "_CPPv4N14torch_tensorrt8DataType8DataTypeE5Value", "torch_tensorrt::DataType::DataType::t"], [0, 3, 1, "_CPPv4N14torch_tensorrt8DataType8DataTypeEN3c1010ScalarTypeE", "torch_tensorrt::DataType::DataType::t"], [0, 4, 1, "_CPPv4N14torch_tensorrt8DataType5ValueE", "torch_tensorrt::DataType::Value"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value5kBoolE", "torch_tensorrt::DataType::Value::kBool"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value5kCharE", "torch_tensorrt::DataType::Value::kChar"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value7kDoubleE", "torch_tensorrt::DataType::Value::kDouble"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value6kFloatE", "torch_tensorrt::DataType::Value::kFloat"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value5kHalfE", "torch_tensorrt::DataType::Value::kHalf"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value4kIntE", "torch_tensorrt::DataType::Value::kInt"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value5kLongE", "torch_tensorrt::DataType::Value::kLong"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value8kUnknownE", "torch_tensorrt::DataType::Value::kUnknown"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value5kBoolE", "torch_tensorrt::DataType::kBool"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value5kCharE", "torch_tensorrt::DataType::kChar"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value7kDoubleE", "torch_tensorrt::DataType::kDouble"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value6kFloatE", "torch_tensorrt::DataType::kFloat"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value5kHalfE", "torch_tensorrt::DataType::kHalf"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value4kIntE", "torch_tensorrt::DataType::kInt"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value5kLongE", "torch_tensorrt::DataType::kLong"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value8kUnknownE", "torch_tensorrt::DataType::kUnknown"], [0, 2, 1, "_CPPv4NK14torch_tensorrt8DataTypecv5ValueEv", "torch_tensorrt::DataType::operator Value"], [0, 2, 1, "_CPPv4N14torch_tensorrt8DataTypecvbEv", "torch_tensorrt::DataType::operator bool"], [0, 2, 1, "_CPPv4NK14torch_tensorrt8DataTypeneE8DataType", "torch_tensorrt::DataType::operator!="], [0, 2, 1, "_CPPv4NK14torch_tensorrt8DataTypeneEN8DataType5ValueE", "torch_tensorrt::DataType::operator!="], [0, 3, 1, "_CPPv4NK14torch_tensorrt8DataTypeneE8DataType", "torch_tensorrt::DataType::operator!=::other"], [0, 3, 1, "_CPPv4NK14torch_tensorrt8DataTypeneEN8DataType5ValueE", "torch_tensorrt::DataType::operator!=::other"], [0, 2, 1, "_CPPv4NK14torch_tensorrt8DataTypeeqE8DataType", "torch_tensorrt::DataType::operator=="], [0, 2, 1, "_CPPv4NK14torch_tensorrt8DataTypeeqEN8DataType5ValueE", "torch_tensorrt::DataType::operator=="], [0, 3, 1, "_CPPv4NK14torch_tensorrt8DataTypeeqE8DataType", "torch_tensorrt::DataType::operator==::other"], [0, 3, 1, "_CPPv4NK14torch_tensorrt8DataTypeeqEN8DataType5ValueE", "torch_tensorrt::DataType::operator==::other"], [46, 1, 1, "_CPPv4N14torch_tensorrt6DeviceE", "torch_tensorrt::Device"], [46, 2, 1, "_CPPv4N14torch_tensorrt6Device6DeviceEv", "torch_tensorrt::Device::Device"], [1, 1, 1, "_CPPv4N14torch_tensorrt6Device10DeviceTypeE", "torch_tensorrt::Device::DeviceType"], [46, 1, 1, "_CPPv4N14torch_tensorrt6Device10DeviceTypeE", "torch_tensorrt::Device::DeviceType"], [1, 2, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeE5Value", "torch_tensorrt::Device::DeviceType::DeviceType"], [1, 2, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeEN3c1010DeviceTypeE", "torch_tensorrt::Device::DeviceType::DeviceType"], [1, 2, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeEv", "torch_tensorrt::Device::DeviceType::DeviceType"], [46, 2, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeE5Value", "torch_tensorrt::Device::DeviceType::DeviceType"], [46, 2, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeEN3c1010DeviceTypeE", "torch_tensorrt::Device::DeviceType::DeviceType"], [46, 2, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeEv", "torch_tensorrt::Device::DeviceType::DeviceType"], [1, 3, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeE5Value", "torch_tensorrt::Device::DeviceType::DeviceType::t"], [1, 3, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeEN3c1010DeviceTypeE", "torch_tensorrt::Device::DeviceType::DeviceType::t"], [46, 3, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeE5Value", "torch_tensorrt::Device::DeviceType::DeviceType::t"], [46, 3, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeEN3c1010DeviceTypeE", "torch_tensorrt::Device::DeviceType::DeviceType::t"], [1, 4, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType5ValueE", "torch_tensorrt::Device::DeviceType::Value"], [46, 4, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType5ValueE", "torch_tensorrt::Device::DeviceType::Value"], [1, 5, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType5Value4kDLAE", "torch_tensorrt::Device::DeviceType::Value::kDLA"], [46, 5, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType5Value4kDLAE", "torch_tensorrt::Device::DeviceType::Value::kDLA"], [1, 5, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType5Value4kGPUE", "torch_tensorrt::Device::DeviceType::Value::kGPU"], [46, 5, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType5Value4kGPUE", "torch_tensorrt::Device::DeviceType::Value::kGPU"], [1, 5, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType5Value4kDLAE", "torch_tensorrt::Device::DeviceType::kDLA"], [1, 5, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType5Value4kGPUE", "torch_tensorrt::Device::DeviceType::kGPU"], [1, 2, 1, "_CPPv4NK14torch_tensorrt6Device10DeviceTypecv5ValueEv", "torch_tensorrt::Device::DeviceType::operator Value"], [46, 2, 1, "_CPPv4NK14torch_tensorrt6Device10DeviceTypecv5ValueEv", "torch_tensorrt::Device::DeviceType::operator Value"], [1, 2, 1, "_CPPv4N14torch_tensorrt6Device10DeviceTypecvbEv", "torch_tensorrt::Device::DeviceType::operator bool"], [46, 2, 1, "_CPPv4N14torch_tensorrt6Device10DeviceTypecvbEv", "torch_tensorrt::Device::DeviceType::operator bool"], [1, 2, 1, "_CPPv4NK14torch_tensorrt6Device10DeviceTypeneE10DeviceType", "torch_tensorrt::Device::DeviceType::operator!="], [46, 2, 1, "_CPPv4NK14torch_tensorrt6Device10DeviceTypeneE10DeviceType", "torch_tensorrt::Device::DeviceType::operator!="], [1, 3, 1, "_CPPv4NK14torch_tensorrt6Device10DeviceTypeneE10DeviceType", "torch_tensorrt::Device::DeviceType::operator!=::other"], [46, 3, 1, "_CPPv4NK14torch_tensorrt6Device10DeviceTypeneE10DeviceType", "torch_tensorrt::Device::DeviceType::operator!=::other"], [1, 2, 1, "_CPPv4NK14torch_tensorrt6Device10DeviceTypeeqE10DeviceType", "torch_tensorrt::Device::DeviceType::operator=="], [46, 2, 1, "_CPPv4NK14torch_tensorrt6Device10DeviceTypeeqE10DeviceType", "torch_tensorrt::Device::DeviceType::operator=="], [1, 3, 1, "_CPPv4NK14torch_tensorrt6Device10DeviceTypeeqE10DeviceType", "torch_tensorrt::Device::DeviceType::operator==::other"], [46, 3, 1, "_CPPv4NK14torch_tensorrt6Device10DeviceTypeeqE10DeviceType", "torch_tensorrt::Device::DeviceType::operator==::other"], [46, 6, 1, "_CPPv4N14torch_tensorrt6Device18allow_gpu_fallbackE", "torch_tensorrt::Device::allow_gpu_fallback"], [46, 6, 1, "_CPPv4N14torch_tensorrt6Device11device_typeE", "torch_tensorrt::Device::device_type"], [46, 6, 1, "_CPPv4N14torch_tensorrt6Device8dla_coreE", "torch_tensorrt::Device::dla_core"], [46, 6, 1, "_CPPv4N14torch_tensorrt6Device6gpu_idE", "torch_tensorrt::Device::gpu_id"], [17, 4, 1, "_CPPv4N14torch_tensorrt16EngineCapabilityE", "torch_tensorrt::EngineCapability"], [17, 5, 1, "_CPPv4N14torch_tensorrt16EngineCapability15kDLA_STANDALONEE", "torch_tensorrt::EngineCapability::kDLA_STANDALONE"], [17, 5, 1, "_CPPv4N14torch_tensorrt16EngineCapability7kSAFETYE", "torch_tensorrt::EngineCapability::kSAFETY"], [17, 5, 1, "_CPPv4N14torch_tensorrt16EngineCapability9kSTANDARDE", "torch_tensorrt::EngineCapability::kSTANDARD"], [47, 1, 1, "_CPPv4N14torch_tensorrt11GraphInputsE", "torch_tensorrt::GraphInputs"], [47, 6, 1, "_CPPv4N14torch_tensorrt11GraphInputs15input_signatureE", "torch_tensorrt::GraphInputs::input_signature"], [47, 6, 1, "_CPPv4N14torch_tensorrt11GraphInputs6inputsE", "torch_tensorrt::GraphInputs::inputs"], [48, 1, 1, "_CPPv4N14torch_tensorrt5InputE", "torch_tensorrt::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputEN2at6TensorE", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputEv", "torch_tensorrt::Input::Input"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::dtype"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::dtype"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::dtype"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::dtype"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::dtype"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::dtype"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::dtype"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::dtype"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input::max_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::max_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::max_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::max_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input::max_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::max_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::max_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::max_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input::min_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::min_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::min_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::min_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input::min_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::min_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::min_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::min_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input::opt_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::opt_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::opt_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::opt_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input::opt_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::opt_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::opt_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::opt_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input::shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input::shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN2at6TensorE", "torch_tensorrt::Input::Input::tensor"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::tensor_domain"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::tensor_domain"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::tensor_domain"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::tensor_domain"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::tensor_domain"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::tensor_domain"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::tensor_domain"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::tensor_domain"], [48, 6, 1, "_CPPv4N14torch_tensorrt5Input5dtypeE", "torch_tensorrt::Input::dtype"], [48, 6, 1, "_CPPv4N14torch_tensorrt5Input6formatE", "torch_tensorrt::Input::format"], [48, 6, 1, "_CPPv4N14torch_tensorrt5Input9max_shapeE", "torch_tensorrt::Input::max_shape"], [48, 6, 1, "_CPPv4N14torch_tensorrt5Input9min_shapeE", "torch_tensorrt::Input::min_shape"], [48, 6, 1, "_CPPv4N14torch_tensorrt5Input9opt_shapeE", "torch_tensorrt::Input::opt_shape"], [48, 6, 1, "_CPPv4N14torch_tensorrt5Input5shapeE", "torch_tensorrt::Input::shape"], [48, 6, 1, "_CPPv4N14torch_tensorrt5Input13tensor_domainE", "torch_tensorrt::Input::tensor_domain"], [2, 1, 1, "_CPPv4N14torch_tensorrt12TensorFormatE", "torch_tensorrt::TensorFormat"], [2, 2, 1, "_CPPv4N14torch_tensorrt12TensorFormat12TensorFormatE5Value", "torch_tensorrt::TensorFormat::TensorFormat"], [2, 2, 1, "_CPPv4N14torch_tensorrt12TensorFormat12TensorFormatEN2at12MemoryFormatE", "torch_tensorrt::TensorFormat::TensorFormat"], [2, 2, 1, "_CPPv4N14torch_tensorrt12TensorFormat12TensorFormatEv", "torch_tensorrt::TensorFormat::TensorFormat"], [2, 3, 1, "_CPPv4N14torch_tensorrt12TensorFormat12TensorFormatE5Value", "torch_tensorrt::TensorFormat::TensorFormat::t"], [2, 3, 1, "_CPPv4N14torch_tensorrt12TensorFormat12TensorFormatEN2at12MemoryFormatE", "torch_tensorrt::TensorFormat::TensorFormat::t"], [2, 4, 1, "_CPPv4N14torch_tensorrt12TensorFormat5ValueE", "torch_tensorrt::TensorFormat::Value"], [2, 5, 1, "_CPPv4N14torch_tensorrt12TensorFormat5Value13kChannelsLastE", "torch_tensorrt::TensorFormat::Value::kChannelsLast"], [2, 5, 1, "_CPPv4N14torch_tensorrt12TensorFormat5Value11kContiguousE", "torch_tensorrt::TensorFormat::Value::kContiguous"], [2, 5, 1, "_CPPv4N14torch_tensorrt12TensorFormat5Value8kUnknownE", "torch_tensorrt::TensorFormat::Value::kUnknown"], [2, 5, 1, "_CPPv4N14torch_tensorrt12TensorFormat5Value13kChannelsLastE", "torch_tensorrt::TensorFormat::kChannelsLast"], [2, 5, 1, "_CPPv4N14torch_tensorrt12TensorFormat5Value11kContiguousE", "torch_tensorrt::TensorFormat::kContiguous"], [2, 5, 1, "_CPPv4N14torch_tensorrt12TensorFormat5Value8kUnknownE", "torch_tensorrt::TensorFormat::kUnknown"], [2, 2, 1, "_CPPv4NK14torch_tensorrt12TensorFormatcv5ValueEv", "torch_tensorrt::TensorFormat::operator Value"], [2, 2, 1, "_CPPv4N14torch_tensorrt12TensorFormatcvbEv", "torch_tensorrt::TensorFormat::operator bool"], [2, 2, 1, "_CPPv4NK14torch_tensorrt12TensorFormatneE12TensorFormat", "torch_tensorrt::TensorFormat::operator!="], [2, 2, 1, "_CPPv4NK14torch_tensorrt12TensorFormatneEN12TensorFormat5ValueE", "torch_tensorrt::TensorFormat::operator!="], [2, 3, 1, "_CPPv4NK14torch_tensorrt12TensorFormatneE12TensorFormat", "torch_tensorrt::TensorFormat::operator!=::other"], [2, 3, 1, "_CPPv4NK14torch_tensorrt12TensorFormatneEN12TensorFormat5ValueE", "torch_tensorrt::TensorFormat::operator!=::other"], [2, 2, 1, "_CPPv4NK14torch_tensorrt12TensorFormateqE12TensorFormat", "torch_tensorrt::TensorFormat::operator=="], [2, 2, 1, "_CPPv4NK14torch_tensorrt12TensorFormateqEN12TensorFormat5ValueE", "torch_tensorrt::TensorFormat::operator=="], [2, 3, 1, "_CPPv4NK14torch_tensorrt12TensorFormateqE12TensorFormat", "torch_tensorrt::TensorFormat::operator==::other"], [2, 3, 1, "_CPPv4NK14torch_tensorrt12TensorFormateqEN12TensorFormat5ValueE", "torch_tensorrt::TensorFormat::operator==::other"], [36, 2, 1, "_CPPv4N14torch_tensorrt15dump_build_infoEv", "torch_tensorrt::dump_build_info"], [34, 2, 1, "_CPPv4N14torch_tensorrt14get_build_infoEv", "torch_tensorrt::get_build_info"], [16, 4, 1, "_CPPv4N14torch_tensorrt7logging5LevelE", "torch_tensorrt::logging::Level"], [16, 5, 1, "_CPPv4N14torch_tensorrt7logging5Level6kDEBUGE", "torch_tensorrt::logging::Level::kDEBUG"], [16, 5, 1, "_CPPv4N14torch_tensorrt7logging5Level6kERRORE", "torch_tensorrt::logging::Level::kERROR"], [16, 5, 1, "_CPPv4N14torch_tensorrt7logging5Level6kGRAPHE", "torch_tensorrt::logging::Level::kGRAPH"], [16, 5, 1, "_CPPv4N14torch_tensorrt7logging5Level5kINFOE", "torch_tensorrt::logging::Level::kINFO"], [16, 5, 1, "_CPPv4N14torch_tensorrt7logging5Level15kINTERNAL_ERRORE", "torch_tensorrt::logging::Level::kINTERNAL_ERROR"], [16, 5, 1, "_CPPv4N14torch_tensorrt7logging5Level8kWARNINGE", "torch_tensorrt::logging::Level::kWARNING"], [24, 2, 1, "_CPPv4N14torch_tensorrt7logging24get_is_colored_output_onEv", "torch_tensorrt::logging::get_is_colored_output_on"], [22, 2, 1, "_CPPv4N14torch_tensorrt7logging18get_logging_prefixEv", "torch_tensorrt::logging::get_logging_prefix"], [23, 2, 1, "_CPPv4N14torch_tensorrt7logging24get_reportable_log_levelEv", "torch_tensorrt::logging::get_reportable_log_level"], [16, 5, 1, "_CPPv4N14torch_tensorrt7logging5Level6kDEBUGE", "torch_tensorrt::logging::kDEBUG"], [16, 5, 1, "_CPPv4N14torch_tensorrt7logging5Level6kERRORE", "torch_tensorrt::logging::kERROR"], [16, 5, 1, "_CPPv4N14torch_tensorrt7logging5Level6kGRAPHE", "torch_tensorrt::logging::kGRAPH"], [16, 5, 1, "_CPPv4N14torch_tensorrt7logging5Level5kINFOE", "torch_tensorrt::logging::kINFO"], [16, 5, 1, "_CPPv4N14torch_tensorrt7logging5Level15kINTERNAL_ERRORE", "torch_tensorrt::logging::kINTERNAL_ERROR"], [16, 5, 1, "_CPPv4N14torch_tensorrt7logging5Level8kWARNINGE", "torch_tensorrt::logging::kWARNING"], [26, 2, 1, "_CPPv4N14torch_tensorrt7logging3logE5LevelNSt6stringE", "torch_tensorrt::logging::log"], [26, 3, 1, "_CPPv4N14torch_tensorrt7logging3logE5LevelNSt6stringE", "torch_tensorrt::logging::log::lvl"], [26, 3, 1, "_CPPv4N14torch_tensorrt7logging3logE5LevelNSt6stringE", "torch_tensorrt::logging::log::msg"], [27, 2, 1, "_CPPv4N14torch_tensorrt7logging24set_is_colored_output_onEb", "torch_tensorrt::logging::set_is_colored_output_on"], [27, 3, 1, "_CPPv4N14torch_tensorrt7logging24set_is_colored_output_onEb", "torch_tensorrt::logging::set_is_colored_output_on::colored_output_on"], [28, 2, 1, "_CPPv4N14torch_tensorrt7logging18set_logging_prefixENSt6stringE", "torch_tensorrt::logging::set_logging_prefix"], [28, 3, 1, "_CPPv4N14torch_tensorrt7logging18set_logging_prefixENSt6stringE", "torch_tensorrt::logging::set_logging_prefix::prefix"], [25, 2, 1, "_CPPv4N14torch_tensorrt7logging24set_reportable_log_levelE5Level", "torch_tensorrt::logging::set_reportable_log_level"], [25, 3, 1, "_CPPv4N14torch_tensorrt7logging24set_reportable_log_levelE5Level", "torch_tensorrt::logging::set_reportable_log_level::lvl"], [3, 1, 1, "_CPPv4I0EN14torch_tensorrt3ptq19Int8CacheCalibratorE", "torch_tensorrt::ptq::Int8CacheCalibrator"], [3, 7, 1, "_CPPv4I0EN14torch_tensorrt3ptq19Int8CacheCalibratorE", "torch_tensorrt::ptq::Int8CacheCalibrator::Algorithm"], [3, 2, 1, "_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibrator19Int8CacheCalibratorERKNSt6stringE", "torch_tensorrt::ptq::Int8CacheCalibrator::Int8CacheCalibrator"], [3, 3, 1, "_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibrator19Int8CacheCalibratorERKNSt6stringE", "torch_tensorrt::ptq::Int8CacheCalibrator::Int8CacheCalibrator::cache_file_path"], [3, 2, 1, "_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibrator8getBatchEA_PvA_PKci", "torch_tensorrt::ptq::Int8CacheCalibrator::getBatch"], [3, 3, 1, "_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibrator8getBatchEA_PvA_PKci", "torch_tensorrt::ptq::Int8CacheCalibrator::getBatch::bindings"], [3, 3, 1, "_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibrator8getBatchEA_PvA_PKci", "torch_tensorrt::ptq::Int8CacheCalibrator::getBatch::names"], [3, 3, 1, "_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibrator8getBatchEA_PvA_PKci", "torch_tensorrt::ptq::Int8CacheCalibrator::getBatch::nbBindings"], [3, 2, 1, "_CPPv4NK14torch_tensorrt3ptq19Int8CacheCalibrator12getBatchSizeEv", "torch_tensorrt::ptq::Int8CacheCalibrator::getBatchSize"], [3, 2, 1, "_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibratorcvPN8nvinfer115IInt8CalibratorEEv", "torch_tensorrt::ptq::Int8CacheCalibrator::operator nvinfer1::IInt8Calibrator*"], [3, 2, 1, "_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibrator20readCalibrationCacheER6size_t", "torch_tensorrt::ptq::Int8CacheCalibrator::readCalibrationCache"], [3, 3, 1, "_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibrator20readCalibrationCacheER6size_t", "torch_tensorrt::ptq::Int8CacheCalibrator::readCalibrationCache::length"], [3, 2, 1, "_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibrator21writeCalibrationCacheEPKv6size_t", "torch_tensorrt::ptq::Int8CacheCalibrator::writeCalibrationCache"], [3, 3, 1, "_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibrator21writeCalibrationCacheEPKv6size_t", "torch_tensorrt::ptq::Int8CacheCalibrator::writeCalibrationCache::cache"], [3, 3, 1, "_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibrator21writeCalibrationCacheEPKv6size_t", "torch_tensorrt::ptq::Int8CacheCalibrator::writeCalibrationCache::length"], [4, 1, 1, "_CPPv4I00EN14torch_tensorrt3ptq14Int8CalibratorE", "torch_tensorrt::ptq::Int8Calibrator"], [4, 7, 1, "_CPPv4I00EN14torch_tensorrt3ptq14Int8CalibratorE", "torch_tensorrt::ptq::Int8Calibrator::Algorithm"], [4, 7, 1, "_CPPv4I00EN14torch_tensorrt3ptq14Int8CalibratorE", "torch_tensorrt::ptq::Int8Calibrator::DataLoaderUniquePtr"], [4, 2, 1, "_CPPv4N14torch_tensorrt3ptq14Int8Calibrator14Int8CalibratorE19DataLoaderUniquePtrRKNSt6stringEb", "torch_tensorrt::ptq::Int8Calibrator::Int8Calibrator"], [4, 3, 1, "_CPPv4N14torch_tensorrt3ptq14Int8Calibrator14Int8CalibratorE19DataLoaderUniquePtrRKNSt6stringEb", "torch_tensorrt::ptq::Int8Calibrator::Int8Calibrator::cache_file_path"], [4, 3, 1, "_CPPv4N14torch_tensorrt3ptq14Int8Calibrator14Int8CalibratorE19DataLoaderUniquePtrRKNSt6stringEb", "torch_tensorrt::ptq::Int8Calibrator::Int8Calibrator::dataloader"], [4, 3, 1, "_CPPv4N14torch_tensorrt3ptq14Int8Calibrator14Int8CalibratorE19DataLoaderUniquePtrRKNSt6stringEb", "torch_tensorrt::ptq::Int8Calibrator::Int8Calibrator::use_cache"], [4, 2, 1, "_CPPv4N14torch_tensorrt3ptq14Int8Calibrator8getBatchEA_PvA_PKci", "torch_tensorrt::ptq::Int8Calibrator::getBatch"], [4, 3, 1, "_CPPv4N14torch_tensorrt3ptq14Int8Calibrator8getBatchEA_PvA_PKci", "torch_tensorrt::ptq::Int8Calibrator::getBatch::bindings"], [4, 3, 1, "_CPPv4N14torch_tensorrt3ptq14Int8Calibrator8getBatchEA_PvA_PKci", "torch_tensorrt::ptq::Int8Calibrator::getBatch::names"], [4, 3, 1, "_CPPv4N14torch_tensorrt3ptq14Int8Calibrator8getBatchEA_PvA_PKci", "torch_tensorrt::ptq::Int8Calibrator::getBatch::nbBindings"], [4, 2, 1, "_CPPv4NK14torch_tensorrt3ptq14Int8Calibrator12getBatchSizeEv", "torch_tensorrt::ptq::Int8Calibrator::getBatchSize"], [4, 2, 1, "_CPPv4N14torch_tensorrt3ptq14Int8CalibratorcvPN8nvinfer115IInt8CalibratorEEv", "torch_tensorrt::ptq::Int8Calibrator::operator nvinfer1::IInt8Calibrator*"], [4, 2, 1, "_CPPv4N14torch_tensorrt3ptq14Int8Calibrator20readCalibrationCacheER6size_t", "torch_tensorrt::ptq::Int8Calibrator::readCalibrationCache"], [4, 3, 1, "_CPPv4N14torch_tensorrt3ptq14Int8Calibrator20readCalibrationCacheER6size_t", "torch_tensorrt::ptq::Int8Calibrator::readCalibrationCache::length"], [4, 2, 1, "_CPPv4N14torch_tensorrt3ptq14Int8Calibrator21writeCalibrationCacheEPKv6size_t", "torch_tensorrt::ptq::Int8Calibrator::writeCalibrationCache"], [4, 3, 1, "_CPPv4N14torch_tensorrt3ptq14Int8Calibrator21writeCalibrationCacheEPKv6size_t", "torch_tensorrt::ptq::Int8Calibrator::writeCalibrationCache::cache"], [4, 3, 1, "_CPPv4N14torch_tensorrt3ptq14Int8Calibrator21writeCalibrationCacheEPKv6size_t", "torch_tensorrt::ptq::Int8Calibrator::writeCalibrationCache::length"], [29, 2, 1, "_CPPv4I0EN14torch_tensorrt3ptq26make_int8_cache_calibratorE19Int8CacheCalibratorI9AlgorithmERKNSt6stringE", "torch_tensorrt::ptq::make_int8_cache_calibrator"], [29, 7, 1, "_CPPv4I0EN14torch_tensorrt3ptq26make_int8_cache_calibratorE19Int8CacheCalibratorI9AlgorithmERKNSt6stringE", "torch_tensorrt::ptq::make_int8_cache_calibrator::Algorithm"], [29, 3, 1, "_CPPv4I0EN14torch_tensorrt3ptq26make_int8_cache_calibratorE19Int8CacheCalibratorI9AlgorithmERKNSt6stringE", "torch_tensorrt::ptq::make_int8_cache_calibrator::cache_file_path"], [30, 2, 1, "_CPPv4I00EN14torch_tensorrt3ptq20make_int8_calibratorE14Int8CalibratorI9Algorithm10DataLoaderE10DataLoaderRKNSt6stringEb", "torch_tensorrt::ptq::make_int8_calibrator"], [30, 7, 1, "_CPPv4I00EN14torch_tensorrt3ptq20make_int8_calibratorE14Int8CalibratorI9Algorithm10DataLoaderE10DataLoaderRKNSt6stringEb", "torch_tensorrt::ptq::make_int8_calibrator::Algorithm"], [30, 7, 1, "_CPPv4I00EN14torch_tensorrt3ptq20make_int8_calibratorE14Int8CalibratorI9Algorithm10DataLoaderE10DataLoaderRKNSt6stringEb", "torch_tensorrt::ptq::make_int8_calibrator::DataLoader"], [30, 3, 1, "_CPPv4I00EN14torch_tensorrt3ptq20make_int8_calibratorE14Int8CalibratorI9Algorithm10DataLoaderE10DataLoaderRKNSt6stringEb", "torch_tensorrt::ptq::make_int8_calibrator::cache_file_path"], [30, 3, 1, "_CPPv4I00EN14torch_tensorrt3ptq20make_int8_calibratorE14Int8CalibratorI9Algorithm10DataLoaderE10DataLoaderRKNSt6stringEb", "torch_tensorrt::ptq::make_int8_calibrator::dataloader"], [30, 3, 1, "_CPPv4I00EN14torch_tensorrt3ptq20make_int8_calibratorE14Int8CalibratorI9Algorithm10DataLoaderE10DataLoaderRKNSt6stringEb", "torch_tensorrt::ptq::make_int8_calibrator::use_cache"], [35, 2, 1, "_CPPv4N14torch_tensorrt10set_deviceEKi", "torch_tensorrt::set_device"], [35, 3, 1, "_CPPv4N14torch_tensorrt10set_deviceEKi", "torch_tensorrt::set_device::gpu_id"], [49, 1, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpecE", "torch_tensorrt::torchscript::CompileSpec"], [49, 2, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecEN5torch3jit6IValueE", "torch_tensorrt::torchscript::CompileSpec::CompileSpec"], [49, 2, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecENSt6vectorI5InputEE", "torch_tensorrt::torchscript::CompileSpec::CompileSpec"], [49, 2, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecENSt6vectorIN3c108ArrayRefI7int64_tEEEE", "torch_tensorrt::torchscript::CompileSpec::CompileSpec"], [49, 2, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecENSt6vectorINSt6vectorI7int64_tEEEE", "torch_tensorrt::torchscript::CompileSpec::CompileSpec"], [49, 3, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecENSt6vectorIN3c108ArrayRefI7int64_tEEEE", "torch_tensorrt::torchscript::CompileSpec::CompileSpec::fixed_sizes"], [49, 3, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecENSt6vectorINSt6vectorI7int64_tEEEE", "torch_tensorrt::torchscript::CompileSpec::CompileSpec::fixed_sizes"], [49, 3, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecEN5torch3jit6IValueE", "torch_tensorrt::torchscript::CompileSpec::CompileSpec::input_signature"], [49, 3, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecENSt6vectorI5InputEE", "torch_tensorrt::torchscript::CompileSpec::CompileSpec::inputs"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec19allow_shape_tensorsE", "torch_tensorrt::torchscript::CompileSpec::allow_shape_tensors"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec10capabilityE", "torch_tensorrt::torchscript::CompileSpec::capability"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec5debugE", "torch_tensorrt::torchscript::CompileSpec::debug"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec6deviceE", "torch_tensorrt::torchscript::CompileSpec::device"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec12disable_tf32E", "torch_tensorrt::torchscript::CompileSpec::disable_tf32"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec20dla_global_dram_sizeE", "torch_tensorrt::torchscript::CompileSpec::dla_global_dram_size"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec19dla_local_dram_sizeE", "torch_tensorrt::torchscript::CompileSpec::dla_local_dram_size"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec13dla_sram_sizeE", "torch_tensorrt::torchscript::CompileSpec::dla_sram_size"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec18enabled_precisionsE", "torch_tensorrt::torchscript::CompileSpec::enabled_precisions"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec12graph_inputsE", "torch_tensorrt::torchscript::CompileSpec::graph_inputs"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec14min_block_sizeE", "torch_tensorrt::torchscript::CompileSpec::min_block_size"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec20num_avg_timing_itersE", "torch_tensorrt::torchscript::CompileSpec::num_avg_timing_iters"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec14ptq_calibratorE", "torch_tensorrt::torchscript::CompileSpec::ptq_calibrator"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec5refitE", "torch_tensorrt::torchscript::CompileSpec::refit"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec24require_full_compilationE", "torch_tensorrt::torchscript::CompileSpec::require_full_compilation"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec14sparse_weightsE", "torch_tensorrt::torchscript::CompileSpec::sparse_weights"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec22torch_executed_modulesE", "torch_tensorrt::torchscript::CompileSpec::torch_executed_modules"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec18torch_executed_opsE", "torch_tensorrt::torchscript::CompileSpec::torch_executed_ops"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec24truncate_long_and_doubleE", "torch_tensorrt::torchscript::CompileSpec::truncate_long_and_double"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec14workspace_sizeE", "torch_tensorrt::torchscript::CompileSpec::workspace_size"], [31, 2, 1, "_CPPv4N14torch_tensorrt11torchscript29check_method_operator_supportERKN5torch3jit6ModuleENSt6stringE", "torch_tensorrt::torchscript::check_method_operator_support"], [31, 3, 1, "_CPPv4N14torch_tensorrt11torchscript29check_method_operator_supportERKN5torch3jit6ModuleENSt6stringE", "torch_tensorrt::torchscript::check_method_operator_support::method_name"], [31, 3, 1, "_CPPv4N14torch_tensorrt11torchscript29check_method_operator_supportERKN5torch3jit6ModuleENSt6stringE", "torch_tensorrt::torchscript::check_method_operator_support::module"], [32, 2, 1, "_CPPv4N14torch_tensorrt11torchscript7compileERKN5torch3jit6ModuleE11CompileSpec", "torch_tensorrt::torchscript::compile"], [32, 3, 1, "_CPPv4N14torch_tensorrt11torchscript7compileERKN5torch3jit6ModuleE11CompileSpec", "torch_tensorrt::torchscript::compile::info"], [32, 3, 1, "_CPPv4N14torch_tensorrt11torchscript7compileERKN5torch3jit6ModuleE11CompileSpec", "torch_tensorrt::torchscript::compile::module"], [37, 2, 1, "_CPPv4N14torch_tensorrt11torchscript28convert_method_to_trt_engineERKN5torch3jit6ModuleENSt6stringE11CompileSpec", "torch_tensorrt::torchscript::convert_method_to_trt_engine"], [37, 3, 1, "_CPPv4N14torch_tensorrt11torchscript28convert_method_to_trt_engineERKN5torch3jit6ModuleENSt6stringE11CompileSpec", "torch_tensorrt::torchscript::convert_method_to_trt_engine::info"], [37, 3, 1, "_CPPv4N14torch_tensorrt11torchscript28convert_method_to_trt_engineERKN5torch3jit6ModuleENSt6stringE11CompileSpec", "torch_tensorrt::torchscript::convert_method_to_trt_engine::method_name"], [37, 3, 1, "_CPPv4N14torch_tensorrt11torchscript28convert_method_to_trt_engineERKN5torch3jit6ModuleENSt6stringE11CompileSpec", "torch_tensorrt::torchscript::convert_method_to_trt_engine::module"], [33, 2, 1, "_CPPv4N14torch_tensorrt11torchscript26embed_engine_in_new_moduleERKNSt6stringE6DeviceRKNSt6vectorINSt6stringEEERKNSt6vectorINSt6stringEEE", "torch_tensorrt::torchscript::embed_engine_in_new_module"], [33, 3, 1, "_CPPv4N14torch_tensorrt11torchscript26embed_engine_in_new_moduleERKNSt6stringE6DeviceRKNSt6vectorINSt6stringEEERKNSt6vectorINSt6stringEEE", "torch_tensorrt::torchscript::embed_engine_in_new_module::device"], [33, 3, 1, "_CPPv4N14torch_tensorrt11torchscript26embed_engine_in_new_moduleERKNSt6stringE6DeviceRKNSt6vectorINSt6stringEEERKNSt6vectorINSt6stringEEE", "torch_tensorrt::torchscript::embed_engine_in_new_module::engine"], [33, 3, 1, "_CPPv4N14torch_tensorrt11torchscript26embed_engine_in_new_moduleERKNSt6stringE6DeviceRKNSt6vectorINSt6stringEEERKNSt6vectorINSt6stringEEE", "torch_tensorrt::torchscript::embed_engine_in_new_module::input_binding_names"], [33, 3, 1, "_CPPv4N14torch_tensorrt11torchscript26embed_engine_in_new_moduleERKNSt6stringE6DeviceRKNSt6vectorINSt6stringEEERKNSt6vectorINSt6stringEEE", "torch_tensorrt::torchscript::embed_engine_in_new_module::output_binding_names"], [76, 8, 0, "-", "torch_tensorrt"]], "torch_tensorrt": [[76, 9, 1, "", "Device"], [76, 9, 1, "", "DeviceType"], [76, 9, 1, "", "EngineCapability"], [76, 9, 1, "", "Input"], [76, 9, 1, "", "MutableTorchTensorRTModule"], [76, 12, 1, "", "compile"], [76, 12, 1, "", "convert_method_to_trt_engine"], [76, 9, 1, "", "dtype"], [121, 8, 0, "-", "dynamo"], [72, 8, 0, "-", "fx"], [76, 12, 1, "", "load"], [73, 8, 0, "-", "logging"], [76, 9, 1, "", "memory_format"], [75, 8, 0, "-", "runtime"], [76, 12, 1, "", "save"], [77, 8, 0, "-", "ts"]], "torch_tensorrt.Device": [[76, 10, 1, "", "__init__"], [76, 11, 1, "", "device_type"], [76, 11, 1, "", "dla_core"], [76, 11, 1, "", "gpu_id"]], "torch_tensorrt.DeviceType": [[76, 11, 1, "", "DLA"], [76, 11, 1, "", "GPU"], [76, 11, 1, "", "UNKNOWN"], [76, 10, 1, "", "to"], [76, 10, 1, "", "try_from"], [76, 10, 1, "", "try_to"]], "torch_tensorrt.EngineCapability": [[76, 11, 1, "", "DLA_STANDALONE"], [76, 11, 1, "", "SAFETY"], [76, 11, 1, "", "STANDARD"], [76, 10, 1, "", "to"], [76, 10, 1, "", "try_from"], [76, 10, 1, "", "try_to"]], "torch_tensorrt.Input": [[76, 10, 1, "", "__init__"], [76, 11, 1, "", "dtype"], [76, 10, 1, "", "example_tensor"], [76, 11, 1, "", "format"], [76, 10, 1, "", "from_tensor"], [76, 10, 1, "", "from_tensors"]], "torch_tensorrt.MutableTorchTensorRTModule": [[76, 10, 1, "", "__init__"], [76, 10, 1, "", "compile"], [76, 10, 1, "", "refit_gm"]], "torch_tensorrt.dtype": [[76, 11, 1, "", "b"], [76, 11, 1, "", "bf16"], [76, 11, 1, "", "f16"], [76, 11, 1, "", "f32"], [76, 11, 1, "", "f64"], [76, 11, 1, "", "f8"], [76, 11, 1, "", "i32"], [76, 11, 1, "", "i64"], [76, 11, 1, "", "i8"], [76, 10, 1, "", "to"], [76, 10, 1, "", "try_from"], [76, 10, 1, "", "try_to"], [76, 11, 1, "", "u8"], [76, 11, 1, "", "unknown"]], "torch_tensorrt.dynamo": [[71, 9, 1, "", "CompilationSettings"], [71, 12, 1, "", "compile"], [71, 12, 1, "", "export"], [71, 12, 1, "", "refit_module_weights"], [71, 12, 1, "", "trace"]], "torch_tensorrt.fx": [[72, 9, 1, "", "InputTensorSpec"], [72, 9, 1, "", "TRTInterpreter"], [72, 9, 1, "", "TRTInterpreterResult"], [72, 9, 1, "", "TRTModule"], [72, 12, 1, "", "compile"]], "torch_tensorrt.logging": [[73, 9, 1, "", "debug"], [73, 9, 1, "", "errors"], [73, 9, 1, "", "graphs"], [73, 9, 1, "", "info"], [73, 9, 1, "", "internal_errors"], [73, 9, 1, "", "warnings"]], "torch_tensorrt.memory_format": [[76, 11, 1, "", "cdhw32"], [76, 11, 1, "", "chw16"], [76, 11, 1, "", "chw2"], [76, 11, 1, "", "chw32"], [76, 11, 1, "", "chw4"], [76, 11, 1, "", "dhwc"], [76, 11, 1, "", "dhwc8"], [76, 11, 1, "", "dla_hwc4"], [76, 11, 1, "", "dla_linear"], [76, 11, 1, "", "hwc"], [76, 11, 1, "", "hwc16"], [76, 11, 1, "", "hwc8"], [76, 11, 1, "", "linear"], [76, 10, 1, "", "to"], [76, 10, 1, "", "try_from"], [76, 10, 1, "", "try_to"]], "torch_tensorrt.runtime": [[75, 9, 1, "", "PythonTorchTensorRTModule"], [75, 9, 1, "", "TorchTensorRTModule"], [75, 12, 1, "", "set_multi_device_safe_mode"]], "torch_tensorrt.runtime.PythonTorchTensorRTModule": [[75, 10, 1, "", "__init__"], [75, 10, 1, "", "disable_profiling"], [75, 10, 1, "", "enable_profiling"], [75, 10, 1, "", "forward"], [75, 10, 1, "", "get_layer_info"], [75, 10, 1, "", "validate_input_shapes"]], "torch_tensorrt.runtime.TorchTensorRTModule": [[75, 10, 1, "", "__init__"], [75, 10, 1, "", "forward"], [75, 10, 1, "", "get_extra_state"], [75, 10, 1, "", "set_extra_state"]], "torch_tensorrt.ts": [[77, 12, 1, "", "TensorRTCompileSpec"], [77, 12, 1, "", "check_method_op_support"], [77, 12, 1, "", "compile"], [77, 12, 1, "", "convert_method_to_trt_engine"], [77, 12, 1, "", "embed_engine_in_new_module"], [74, 8, 0, "-", "ptq"]], "torch_tensorrt.ts.ptq": [[74, 9, 1, "", "CacheCalibrator"], [74, 9, 1, "", "CalibrationAlgo"], [74, 9, 1, "", "DataLoaderCalibrator"]], "torch_tensorrt.ts.ptq.CalibrationAlgo": [[74, 11, 1, "", "ENTROPY_CALIBRATION"], [74, 11, 1, "", "ENTROPY_CALIBRATION_2"], [74, 11, 1, "", "LEGACY_CALIBRATION"], [74, 11, 1, "", "MINMAX_CALIBRATION"]]}, "objtypes": {"0": "c:macro", "1": "cpp:class", "2": "cpp:function", "3": "cpp:functionParam", "4": "cpp:enum", "5": "cpp:enumerator", "6": "cpp:member", "7": "cpp:templateParam", "8": "py:module", "9": "py:class", "10": "py:method", "11": "py:attribute", "12": "py:function"}, "objnames": {"0": ["c", "macro", "C macro"], "1": ["cpp", "class", "C++ class"], "2": ["cpp", "function", "C++ function"], "3": ["cpp", "functionParam", "C++ function parameter"], "4": ["cpp", "enum", "C++ enum"], "5": ["cpp", "enumerator", "C++ enumerator"], "6": ["cpp", "member", "C++ member"], "7": ["cpp", "templateParam", "C++ template parameter"], "8": ["py", "module", "Python module"], "9": ["py", "class", "Python class"], "10": ["py", "method", "Python method"], "11": ["py", "attribute", "Python attribute"], "12": ["py", "function", "Python function"]}, "titleterms": {"class": [0, 1, 2, 3, 4, 20, 21, 38, 40, 41, 50, 71, 72, 74, 75, 76], "datatyp": 0, "document": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 16, 17, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 46, 47, 48, 49, 61, 69, 85, 86], "devic": [1, 46, 120], "devicetyp": 1, "nest": [1, 46], "relationship": [1, 3, 4, 46, 48], "tensorformat": 2, "templat": [3, 4, 29, 30], "int8cachecalibr": 3, "inherit": [3, 4, 48], "base": [3, 4, 48, 80], "type": [3, 4, 46, 48, 54], "int8calibr": 4, "defin": [5, 6, 7, 8, 9, 10, 11, 12, 19, 50, 101, 104, 111, 112], "str": 5, "torch_tensorrt_patch_vers": 6, "torch_tensorrt_major_vers": 7, "torch_tensorrt_minor_vers": 8, "torchtrt_api": 9, "xstr": 10, "torchtrt_hidden": 11, "torch_tensorrt_vers": 12, "directori": [13, 14, 15, 51], "cpp": [13, 18, 19, 20, 21, 56], "subdirectori": [13, 14], "includ": [14, 18, 19, 20, 21], "torch_tensorrt": [15, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 45, 67, 71, 72, 73, 74, 75, 76, 77, 105, 107, 108, 122], "file": [15, 18, 19, 20, 21, 42, 43, 44, 45, 50, 51], "enum": [16, 17, 18, 21, 38, 39, 50, 74, 76], "level": [16, 80, 82, 83], "enginecap": 17, "log": [18, 22, 23, 24, 25, 26, 27, 28, 39, 42, 73], "h": [18, 19, 20, 21, 42, 43, 44, 45, 56], "content": [18, 19, 20, 21, 38, 39, 40, 41, 80, 81, 82, 83, 84, 85], "definit": [18, 19, 20, 21, 83, 95, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 112, 113], "By": [18, 19], "namespac": [18, 19, 20, 21, 38, 39, 40, 41, 50], "function": [18, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 50, 61, 71, 72, 75, 76, 77, 101, 112], "macro": [19, 43], "ptq": [20, 29, 30, 40, 44, 74, 91, 112], "get_logging_prefix": 22, "get_reportable_log_level": 23, "get_is_colored_output_on": 24, "set_reportable_log_level": 25, "set_is_colored_output_on": 27, "set_logging_prefix": 28, "make_int8_cache_calibr": 29, "make_int8_calibr": 30, "torchscript": [31, 32, 33, 37, 41, 60, 66, 69, 88, 89, 92, 121, 122], "check_method_operator_support": 31, "compil": [32, 57, 59, 63, 64, 66, 68, 69, 89, 95, 98, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 113, 116, 118, 119, 121, 122], "embed_engine_in_new_modul": 33, "get_build_info": 34, "set_devic": 35, "dump_build_info": 36, "convert_method_to_trt_engin": 37, "program": [42, 43, 44, 45, 63, 102, 120], "list": [42, 43, 44, 45, 83], "struct": [46, 47, 48, 49, 50], "graphinput": 47, "input": [48, 105, 107, 111], "compilespec": 49, "torch": [50, 61, 63, 64, 65, 66, 68, 69, 89, 90, 92, 94, 96, 100, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 114, 115, 116, 117, 118, 119, 120, 121, 122], "tensorrt": [50, 58, 61, 63, 64, 65, 66, 69, 89, 90, 92, 93, 94, 96, 100, 102, 104, 109, 110, 111, 112, 114, 115, 116, 117, 118, 119, 120, 121, 122], "c": [50, 61, 66, 68, 69, 89, 91, 116], "api": [50, 51, 61, 66, 69, 101], "hierarchi": 50, "full": [50, 51], "torchtrtc": [52, 89], "convers": [53, 57, 59, 60], "phase": [53, 55, 56, 57, 58, 59], "node": 53, "evalu": [53, 54, 70], "convert": [53, 54, 60, 65, 70, 89, 93, 94], "write": [54, 60, 62, 93, 94, 96], "dynamo": [54, 62, 69, 71, 109, 110, 111, 121, 122], "implement": [54, 94], "registr": 54, "capabl": 54, "valid": 54, "contract": [54, 60], "exampl": [54, 62, 82, 84, 95], "convolut": 54, "oper": [54, 64, 70, 89, 93, 96], "decomposit": 54, "addmm": [54, 55], "lower": [55, 57, 59, 62], "pass": [55, 62], "us": [55, 61, 89, 90, 92, 93, 94, 96, 101, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 116, 118], "eliminatecommonsubexpress": 55, "elimin": 55, "dead": 55, "code": [55, 69, 82], "except": 55, "Or": 55, "pattern": 55, "redund": 55, "guard": 55, "freez": 55, "modul": [55, 88, 89, 100, 108, 122], "fuse": 55, "branch": 55, "linear": 55, "flatten": 55, "graph": [55, 58, 108, 122], "tupl": 55, "fallback": [55, 56], "peephol": 55, "optim": [55, 68, 114, 115, 117], "remov": 55, "contigu": 55, "dropout": 55, "To": 55, "unpack": 55, "logsoftmax": 55, "unrol": 55, "loop": [55, 112], "replac": [55, 82], "tile": 55, "repeat": 55, "partit": [56, 57, 59], "partitoninfo": 56, "segmentedblock": 56, "shape_analysi": 56, "automat": [56, 93, 113], "depend": [56, 66, 99, 114], "awar": [56, 116], "runtim": [57, 58, 59, 75, 95, 101, 120], "background": [58, 60], "engin": [58, 65, 96, 97, 98], "executor": 58, "op": [58, 65, 96], "construct": 58, "result": 58, "serial": [58, 64, 68], "deseri": 58, "abi": [58, 66], "version": [58, 66], "format": [58, 122], "system": [59, 66, 93], "overview": [59, 67], "what": 60, "guarante": 60, "respons": 60, "context": [60, 80, 113], "arg": [60, 81], "weight": [60, 102, 111, 112, 113], "other": 60, "advic": 60, "link": [61, 82], "develop": 61, "avail": 61, "layer": 61, "expect": 61, "dimens": 61, "python": [61, 66, 68, 69, 88, 90, 91], "sometim": 61, "easier": 61, "read": 61, "pytorch": [61, 65, 69, 92, 93, 96, 104, 109, 110, 116], "native_op": 61, "ir": [61, 121, 122], "aten": 62, "basic": 62, "requir": 62, "regist": [62, 89], "export": [63, 68, 108, 118], "customiz": [63, 64], "set": [63, 64, 100, 103, 108, 114, 115, 117], "under": [63, 89, 118], "hood": [63, 89, 118], "trace": 63, "backend": [64, 105, 106, 107, 109, 110, 111], "kei": 64, "featur": [64, 101], "custom": [64, 89, 93, 94, 96, 98, 103, 118], "usag": [64, 102, 103], "after": 64, "model": [64, 65, 69, 93, 95, 96, 99, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 119, 121], "perform": [64, 101], "coverag": 64, "feasibl": 64, "dynam": [64, 105, 116, 118], "shape": [64, 105, 116, 118], "support": [64, 70], "recompil": [64, 105], "condit": 64, "fx": [65, 69, 72, 116, 122], "frontend": [65, 66, 69, 92, 104, 116, 122], "user": [65, 69], "guid": [65, 69], "acc": 65, "tracer": 65, "fx2trt": 65, "how": [65, 80, 91], "add": 65, "miss": 65, "instal": [66, 87], "precompil": 66, "binari": 66, "specif": 66, "cuda": [66, 103, 107, 108], "nightli": 66, "build": [66, 67, 80, 114, 115, 117], "onli": 66, "from": [66, 92], "sourc": 66, "linux": 66, "packag": [66, 120], "addit": 66, "option": [66, 68, 80, 81, 83, 105, 107, 113, 122], "distribut": 66, "No": 66, "librari": [66, 104, 111, 120], "standalon": 66, "releas": 66, "debug": 66, "pre": [66, 101, 112], "cxx11": 66, "choos": 66, "right": 66, "window": [66, 95], "step": [66, 68, 114, 115, 117], "advanc": [66, 102, 103], "setup": 66, "troubleshoot": 66, "altern": 66, "cmake": 66, "nativ": 66, "aarch64": 66, "jetson": 66, "prerequisit": [66, 67], "environ": 66, "cli": [66, 69], "jetpack": 67, "6": [67, 84], "1": [67, 68, 84, 114, 115, 117], "quick": [68, 93], "start": [68, 69], "2": [68, 84, 85, 114, 115, 117], "deploi": [68, 93, 112, 116, 120], "deploy": 68, "In": [69, 102], "framework": 69, "infer": [69, 101, 104, 105, 106, 107, 108, 112, 114, 115, 117], "nvidia": 69, "gpu": 69, "get": 69, "tutori": [69, 114], "zoo": [69, 99, 114], "contributor": 69, "indic": 69, "legaci": [69, 116, 122], "further": 69, "inform": 69, "current": 70, "through": 70, "ts": [74, 77, 122], "submodul": 76, "comput": 78, "time": [78, 122], "changelog": 79, "configur": 80, "project": 80, "wide": 80, "html": 80, "theme": [80, 86], "toc": 80, "page": 80, "tabl": [80, 81, 82, 83, 84, 85], "mod": 81, "test_py_modul": 81, "gener": [81, 93, 109, 110], "index": 81, "paramet": [81, 104], "data": 81, "paragraph": [82, 85], "markup": 82, "inlin": 82, "math": 82, "meta": 82, "block": 82, "liter": 82, "line": 82, "quot": 82, "doctest": 82, "emphas": 82, "number": [82, 83], "sidebar": 82, "ch": 82, "ien": 82, "The": [82, 89], "creativ": 82, "A": 82, "refer": [82, 111], "footnot": 82, "citat": [82, 91], "glossari": 82, "target": 82, "direct": 82, "center": 82, "text": 82, "imag": [82, 83, 111], "figur": 82, "admonit": 82, "And": 82, "wai": 82, "topic": 82, "rubric": 82, "titl": 82, "compound": 82, "download": [82, 87], "enumer": 83, "field": 83, "bullet": 83, "second": 83, "But": 83, "deeper": 83, "down": 83, "rabbit": 83, "hole": 83, "hlist": 83, "grid": 83, "giant": 83, "can": 83, "have": 83, "caption": [83, 86], "like": 83, "thi": [83, 86], "one": 83, "long": [84, 86], "sticki": 84, "nav": 84, "menu": [84, 86], "3": [84, 114, 115, 117], "4": 84, "5": 84, "7": 84, "8": 84, "9": 84, "10": 84, "11": 84, "12": 84, "13": 84, "14": 84, "15": 84, "16": 84, "17": 84, "18": 84, "19": 84, "20": 84, "submenu": 84, "subsubmenu": 84, "structur": 85, "element": 85, "section": 85, "subsect": 85, "subsubsect": 85, "demo": 86, "an": [86, 111], "incred": 86, "via": 87, "git": 87, "creat": [88, 91], "work": [88, 89], "save": [88, 100, 121], "disk": 88, "quickstart": 89, "unsupport": 89, "post": [91, 111], "train": [91, 112, 116], "quantiz": [91, 112, 116], "your": [91, 114, 115, 117], "own": 91, "applic": 91, "directli": 92, "kernel": [93, 96], "plugin": [93, 120], "our": [93, 94, 96], "overload": 94, "metadata": 94, "cross": 95, "import": [95, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113], "within": 96, "test": 96, "wrap": 96, "insert": 96, "cach": [97, 98, 102], "bert": [97, 107, 116], "jit": [98, 118], "aot": [98, 118], "mutabl": 100, "initi": [100, 111], "make": [100, 102], "modif": 100, "stabl": [100, 106], "diffus": [100, 106], "huggingfac": 100, "alloc": 101, "output": [101, 104, 109, 110, 111], "buffer": 101, "measur": 101, "load": [101, 111, 112, 121], "enabl": 101, "disabl": 101, "refit": 102, "new": 102, "standard": 102, "workflow": 102, "refitt": 102, "pretrain": [102, 111], "map": 102, "place": 102, "default": [103, 108], "cleanup": [103, 107], "driver": [103, 107], "error": [103, 107], "note": [103, 107], "gpt2": [104, 109], "necessari": 104, "decod": [104, 109, 110], "sentenc": [104, 109, 110], "resnet": 105, "argument": [105, 107], "avoid": 105, "specifi": 105, "befor": 105, "trt": 105, "cudagraph": [108, 120], "integr": 108, "contain": 108, "break": 108, "llama2": 110, "sam2": 111, "follow": 111, "preprocess": 111, "compon": 111, "process": 111, "visual": 111, "dataset": 112, "loss": 112, "calibr": 112, "tune": 112, "fp8": 112, "stream": 113, "run": 113, "budget": 113, "size": 113, "manag": 113, "serv": [114, 115, 116, 117], "triton": [114, 115, 117], "up": [114, 115, 117], "server": [114, 115, 117], "client": [114, 115, 117], "queri": [114, 115, 117], "notebook": 116, "citrinet": 116, "efficientnet": 116, "mask": 116, "languag": 116, "mlm": 116, "hug": 116, "face": 116, "transform": 116, "acceler": 116, "resnet50": 116, "lenet": 116, "deep": 116, "learn": 116, "object": 116, "detect": 116, "ssd": 116, "int8": 116, "constraint": 118, "mix": 119, "precis": 119, "libtorchtrt": 120, "so": 120, "multi": 120, "safe": 120, "mode": 120, "exportedprogram": 121, "b": 121, "explain": 122, "just": 122, "accept": 122, "return": 122, "ahead": 122, "dla": 123}, "envversion": {"sphinx.domains.c": 2, "sphinx.domains.changeset": 1, "sphinx.domains.citation": 1, "sphinx.domains.cpp": 6, "sphinx.domains.index": 1, "sphinx.domains.javascript": 2, "sphinx.domains.math": 2, "sphinx.domains.python": 3, "sphinx.domains.rst": 2, "sphinx.domains.std": 2, "nbsphinx": 4, "sphinx.ext.intersphinx": 1, "sphinx.ext.todo": 2, "sphinx.ext.viewcode": 1, "sphinx": 56}}) \ No newline at end of file +Search.setIndex({"docnames": ["_cpp_api/classtorch__tensorrt_1_1DataType", "_cpp_api/classtorch__tensorrt_1_1Device_1_1DeviceType", "_cpp_api/classtorch__tensorrt_1_1TensorFormat", "_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8CacheCalibrator", "_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8Calibrator", "_cpp_api/define_macros_8h_1a18d295a837ac71add5578860b55e5502", "_cpp_api/define_macros_8h_1a282fd3c0b1c3a215148ae372070e1268", "_cpp_api/define_macros_8h_1a31398a6d4d27e28817afb0f0139e909e", "_cpp_api/define_macros_8h_1a35703561b26b1a9d2738ad7d58b27827", "_cpp_api/define_macros_8h_1abd1465eb38256d3f22cc1426b23d516b", "_cpp_api/define_macros_8h_1abe87b341f562fd1cf40b7672e4d759da", "_cpp_api/define_macros_8h_1ad19939408f7be171a74a89928b36eb59", "_cpp_api/define_macros_8h_1adad592a7b1b7eed529cdf6acd584c883", "_cpp_api/dir_cpp", "_cpp_api/dir_cpp_include", "_cpp_api/dir_cpp_include_torch_tensorrt", "_cpp_api/enum_logging_8h_1a130f65408ad8cbaee060f05e8db69558", "_cpp_api/enum_torch__tensorrt_8h_1a3fbe5d72e4fc624dbd038853079620eb", "_cpp_api/file_cpp_include_torch_tensorrt_logging.h", "_cpp_api/file_cpp_include_torch_tensorrt_macros.h", "_cpp_api/file_cpp_include_torch_tensorrt_ptq.h", "_cpp_api/file_cpp_include_torch_tensorrt_torch_tensorrt.h", "_cpp_api/function_logging_8h_1a0593f776f469c20469e2f729fc7861a3", "_cpp_api/function_logging_8h_1a0c012cb374addd90eb1f42eaec570650", "_cpp_api/function_logging_8h_1a56e110feaaba2c3fd44bd201fd21a76a", "_cpp_api/function_logging_8h_1a7cb50492421ea9de4e3db895819df6f2", "_cpp_api/function_logging_8h_1ac46ac0901cb97e3ae6e93b45f24e90b8", "_cpp_api/function_logging_8h_1ad2efd47b6c3689e58ccc595680579ae5", "_cpp_api/function_logging_8h_1af8f3443813315af7901903d25dd495cc", "_cpp_api/function_ptq_8h_1a226e3c83379d1012cde8578c1c86b16c", "_cpp_api/function_ptq_8h_1a6186e305f47c1d94b6130ef6c7f7e178", "_cpp_api/function_torch__tensorrt_8h_1a5b405fd3bf3c8fc2e2a54cbbab979797", "_cpp_api/function_torch__tensorrt_8h_1a6e19490a08fb1553c9dd347a5ae79db9", "_cpp_api/function_torch__tensorrt_8h_1a81f9783517335dda877d8cfcf38987c9", "_cpp_api/function_torch__tensorrt_8h_1ac4ab8313ae72c2c899ea31548b528528", "_cpp_api/function_torch__tensorrt_8h_1ad1acd06eaeaffbbcf6e7ebf426891384", "_cpp_api/function_torch__tensorrt_8h_1ad6a4ee8ca6c8f6e5519eb1128ec7f4a1", "_cpp_api/function_torch__tensorrt_8h_1ae8d56472106eeef37fbe51ff7f40c9b2", "_cpp_api/namespace_torch_tensorrt", "_cpp_api/namespace_torch_tensorrt__logging", "_cpp_api/namespace_torch_tensorrt__ptq", "_cpp_api/namespace_torch_tensorrt__torchscript", "_cpp_api/program_listing_file_cpp_include_torch_tensorrt_logging.h", "_cpp_api/program_listing_file_cpp_include_torch_tensorrt_macros.h", "_cpp_api/program_listing_file_cpp_include_torch_tensorrt_ptq.h", "_cpp_api/program_listing_file_cpp_include_torch_tensorrt_torch_tensorrt.h", "_cpp_api/structtorch__tensorrt_1_1Device", "_cpp_api/structtorch__tensorrt_1_1GraphInputs", "_cpp_api/structtorch__tensorrt_1_1Input", "_cpp_api/structtorch__tensorrt_1_1torchscript_1_1CompileSpec", "_cpp_api/torch_tensort_cpp", "_cpp_api/unabridged_orphan", "cli/torchtrtc", "contributors/conversion", "contributors/dynamo_converters", "contributors/lowering", "contributors/partitioning", "contributors/phases", "contributors/runtime", "contributors/system_overview", "contributors/ts_converters", "contributors/useful_links", "contributors/writing_dynamo_aten_lowering_passes", "dynamo/dynamo_export", "dynamo/torch_compile", "fx/getting_started_with_fx_path", "getting_started/installation", "getting_started/jetpack", "getting_started/quick_start", "index", "indices/supported_ops", "py_api/dynamo", "py_api/fx", "py_api/logging", "py_api/ptq", "py_api/runtime", "py_api/torch_tensorrt", "py_api/ts", "sg_execution_times", "src/pytorch-sphinx-theme/docs/changelog", "src/pytorch-sphinx-theme/docs/configuring", "src/pytorch-sphinx-theme/docs/demo/api", "src/pytorch-sphinx-theme/docs/demo/demo", "src/pytorch-sphinx-theme/docs/demo/lists_tables", "src/pytorch-sphinx-theme/docs/demo/long", "src/pytorch-sphinx-theme/docs/demo/structure", "src/pytorch-sphinx-theme/docs/index", "src/pytorch-sphinx-theme/docs/installing", "ts/creating_torchscript_module_in_python", "ts/getting_started_with_cpp_api", "ts/getting_started_with_python_api", "ts/ptq", "ts/torchscript_frontend_from_pytorch", "tutorials/_rendered_examples/dynamo/auto_generate_converters", "tutorials/_rendered_examples/dynamo/converter_overloading", "tutorials/_rendered_examples/dynamo/cross_runtime_compilation_for_windows", "tutorials/_rendered_examples/dynamo/custom_kernel_plugins", "tutorials/_rendered_examples/dynamo/engine_caching_bert_example", "tutorials/_rendered_examples/dynamo/engine_caching_example", "tutorials/_rendered_examples/dynamo/index", "tutorials/_rendered_examples/dynamo/mutable_torchtrt_module_example", "tutorials/_rendered_examples/dynamo/pre_allocated_output_example", "tutorials/_rendered_examples/dynamo/refit_engine_example", "tutorials/_rendered_examples/dynamo/torch_compile_advanced_usage", "tutorials/_rendered_examples/dynamo/torch_compile_gpt2", "tutorials/_rendered_examples/dynamo/torch_compile_resnet_example", "tutorials/_rendered_examples/dynamo/torch_compile_stable_diffusion", "tutorials/_rendered_examples/dynamo/torch_compile_transformers_example", "tutorials/_rendered_examples/dynamo/torch_export_cudagraphs", "tutorials/_rendered_examples/dynamo/torch_export_gpt2", "tutorials/_rendered_examples/dynamo/torch_export_llama2", "tutorials/_rendered_examples/dynamo/torch_export_sam2", "tutorials/_rendered_examples/dynamo/vgg16_ptq", "tutorials/_rendered_examples/dynamo/weight_streaming_example", "tutorials/_rendered_examples/index", "tutorials/_rendered_examples/triton/index", "tutorials/notebooks", "tutorials/serving_torch_tensorrt_with_triton", "user_guide/dynamic_shapes", "user_guide/mixed_precision", "user_guide/runtime", "user_guide/saving_models", "user_guide/torch_tensorrt_explained", "user_guide/using_dla"], "filenames": ["_cpp_api/classtorch__tensorrt_1_1DataType.rst", "_cpp_api/classtorch__tensorrt_1_1Device_1_1DeviceType.rst", "_cpp_api/classtorch__tensorrt_1_1TensorFormat.rst", "_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8CacheCalibrator.rst", "_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8Calibrator.rst", "_cpp_api/define_macros_8h_1a18d295a837ac71add5578860b55e5502.rst", "_cpp_api/define_macros_8h_1a282fd3c0b1c3a215148ae372070e1268.rst", "_cpp_api/define_macros_8h_1a31398a6d4d27e28817afb0f0139e909e.rst", "_cpp_api/define_macros_8h_1a35703561b26b1a9d2738ad7d58b27827.rst", "_cpp_api/define_macros_8h_1abd1465eb38256d3f22cc1426b23d516b.rst", "_cpp_api/define_macros_8h_1abe87b341f562fd1cf40b7672e4d759da.rst", "_cpp_api/define_macros_8h_1ad19939408f7be171a74a89928b36eb59.rst", "_cpp_api/define_macros_8h_1adad592a7b1b7eed529cdf6acd584c883.rst", "_cpp_api/dir_cpp.rst", "_cpp_api/dir_cpp_include.rst", "_cpp_api/dir_cpp_include_torch_tensorrt.rst", "_cpp_api/enum_logging_8h_1a130f65408ad8cbaee060f05e8db69558.rst", "_cpp_api/enum_torch__tensorrt_8h_1a3fbe5d72e4fc624dbd038853079620eb.rst", "_cpp_api/file_cpp_include_torch_tensorrt_logging.h.rst", "_cpp_api/file_cpp_include_torch_tensorrt_macros.h.rst", "_cpp_api/file_cpp_include_torch_tensorrt_ptq.h.rst", "_cpp_api/file_cpp_include_torch_tensorrt_torch_tensorrt.h.rst", "_cpp_api/function_logging_8h_1a0593f776f469c20469e2f729fc7861a3.rst", "_cpp_api/function_logging_8h_1a0c012cb374addd90eb1f42eaec570650.rst", "_cpp_api/function_logging_8h_1a56e110feaaba2c3fd44bd201fd21a76a.rst", "_cpp_api/function_logging_8h_1a7cb50492421ea9de4e3db895819df6f2.rst", "_cpp_api/function_logging_8h_1ac46ac0901cb97e3ae6e93b45f24e90b8.rst", "_cpp_api/function_logging_8h_1ad2efd47b6c3689e58ccc595680579ae5.rst", "_cpp_api/function_logging_8h_1af8f3443813315af7901903d25dd495cc.rst", "_cpp_api/function_ptq_8h_1a226e3c83379d1012cde8578c1c86b16c.rst", "_cpp_api/function_ptq_8h_1a6186e305f47c1d94b6130ef6c7f7e178.rst", "_cpp_api/function_torch__tensorrt_8h_1a5b405fd3bf3c8fc2e2a54cbbab979797.rst", "_cpp_api/function_torch__tensorrt_8h_1a6e19490a08fb1553c9dd347a5ae79db9.rst", "_cpp_api/function_torch__tensorrt_8h_1a81f9783517335dda877d8cfcf38987c9.rst", "_cpp_api/function_torch__tensorrt_8h_1ac4ab8313ae72c2c899ea31548b528528.rst", "_cpp_api/function_torch__tensorrt_8h_1ad1acd06eaeaffbbcf6e7ebf426891384.rst", "_cpp_api/function_torch__tensorrt_8h_1ad6a4ee8ca6c8f6e5519eb1128ec7f4a1.rst", "_cpp_api/function_torch__tensorrt_8h_1ae8d56472106eeef37fbe51ff7f40c9b2.rst", "_cpp_api/namespace_torch_tensorrt.rst", "_cpp_api/namespace_torch_tensorrt__logging.rst", "_cpp_api/namespace_torch_tensorrt__ptq.rst", "_cpp_api/namespace_torch_tensorrt__torchscript.rst", "_cpp_api/program_listing_file_cpp_include_torch_tensorrt_logging.h.rst", "_cpp_api/program_listing_file_cpp_include_torch_tensorrt_macros.h.rst", "_cpp_api/program_listing_file_cpp_include_torch_tensorrt_ptq.h.rst", "_cpp_api/program_listing_file_cpp_include_torch_tensorrt_torch_tensorrt.h.rst", "_cpp_api/structtorch__tensorrt_1_1Device.rst", "_cpp_api/structtorch__tensorrt_1_1GraphInputs.rst", "_cpp_api/structtorch__tensorrt_1_1Input.rst", "_cpp_api/structtorch__tensorrt_1_1torchscript_1_1CompileSpec.rst", "_cpp_api/torch_tensort_cpp.rst", "_cpp_api/unabridged_orphan.rst", "cli/torchtrtc.rst", "contributors/conversion.rst", "contributors/dynamo_converters.rst", "contributors/lowering.rst", "contributors/partitioning.rst", "contributors/phases.rst", "contributors/runtime.rst", "contributors/system_overview.rst", "contributors/ts_converters.rst", "contributors/useful_links.rst", "contributors/writing_dynamo_aten_lowering_passes.rst", "dynamo/dynamo_export.rst", "dynamo/torch_compile.rst", "fx/getting_started_with_fx_path.rst", "getting_started/installation.rst", "getting_started/jetpack.rst", "getting_started/quick_start.rst", "index.rst", "indices/supported_ops.rst", "py_api/dynamo.rst", "py_api/fx.rst", "py_api/logging.rst", "py_api/ptq.rst", "py_api/runtime.rst", "py_api/torch_tensorrt.rst", "py_api/ts.rst", "sg_execution_times.rst", "src/pytorch-sphinx-theme/docs/changelog.rst", "src/pytorch-sphinx-theme/docs/configuring.rst", "src/pytorch-sphinx-theme/docs/demo/api.rst", "src/pytorch-sphinx-theme/docs/demo/demo.rst", "src/pytorch-sphinx-theme/docs/demo/lists_tables.rst", "src/pytorch-sphinx-theme/docs/demo/long.rst", "src/pytorch-sphinx-theme/docs/demo/structure.rst", "src/pytorch-sphinx-theme/docs/index.rst", "src/pytorch-sphinx-theme/docs/installing.rst", "ts/creating_torchscript_module_in_python.rst", "ts/getting_started_with_cpp_api.rst", "ts/getting_started_with_python_api.rst", "ts/ptq.rst", "ts/torchscript_frontend_from_pytorch.rst", "tutorials/_rendered_examples/dynamo/auto_generate_converters.rst", "tutorials/_rendered_examples/dynamo/converter_overloading.rst", "tutorials/_rendered_examples/dynamo/cross_runtime_compilation_for_windows.rst", "tutorials/_rendered_examples/dynamo/custom_kernel_plugins.rst", "tutorials/_rendered_examples/dynamo/engine_caching_bert_example.rst", "tutorials/_rendered_examples/dynamo/engine_caching_example.rst", "tutorials/_rendered_examples/dynamo/index.rst", "tutorials/_rendered_examples/dynamo/mutable_torchtrt_module_example.rst", "tutorials/_rendered_examples/dynamo/pre_allocated_output_example.rst", "tutorials/_rendered_examples/dynamo/refit_engine_example.rst", "tutorials/_rendered_examples/dynamo/torch_compile_advanced_usage.rst", "tutorials/_rendered_examples/dynamo/torch_compile_gpt2.rst", "tutorials/_rendered_examples/dynamo/torch_compile_resnet_example.rst", "tutorials/_rendered_examples/dynamo/torch_compile_stable_diffusion.rst", "tutorials/_rendered_examples/dynamo/torch_compile_transformers_example.rst", "tutorials/_rendered_examples/dynamo/torch_export_cudagraphs.rst", "tutorials/_rendered_examples/dynamo/torch_export_gpt2.rst", "tutorials/_rendered_examples/dynamo/torch_export_llama2.rst", "tutorials/_rendered_examples/dynamo/torch_export_sam2.rst", "tutorials/_rendered_examples/dynamo/vgg16_ptq.rst", "tutorials/_rendered_examples/dynamo/weight_streaming_example.rst", "tutorials/_rendered_examples/index.rst", "tutorials/_rendered_examples/triton/index.rst", "tutorials/notebooks.rst", "tutorials/serving_torch_tensorrt_with_triton.rst", "user_guide/dynamic_shapes.rst", "user_guide/mixed_precision.rst", "user_guide/runtime.rst", "user_guide/saving_models.rst", "user_guide/torch_tensorrt_explained.rst", "user_guide/using_dla.rst"], "titles": ["Class DataType", "Class Device::DeviceType", "Class TensorFormat", "Template Class Int8CacheCalibrator", "Template Class Int8Calibrator", "Define STR", "Define TORCH_TENSORRT_PATCH_VERSION", "Define TORCH_TENSORRT_MAJOR_VERSION", "Define TORCH_TENSORRT_MINOR_VERSION", "Define TORCHTRT_API", "Define XSTR", "Define TORCHTRT_HIDDEN", "Define TORCH_TENSORRT_VERSION", "Directory cpp", "Directory include", "Directory torch_tensorrt", "Enum Level", "Enum EngineCapability", "File logging.h", "File macros.h", "File ptq.h", "File torch_tensorrt.h", "Function torch_tensorrt::logging::get_logging_prefix", "Function torch_tensorrt::logging::get_reportable_log_level", "Function torch_tensorrt::logging::get_is_colored_output_on", "Function torch_tensorrt::logging::set_reportable_log_level", "Function torch_tensorrt::logging::log", "Function torch_tensorrt::logging::set_is_colored_output_on", "Function torch_tensorrt::logging::set_logging_prefix", "Template Function torch_tensorrt::ptq::make_int8_cache_calibrator", "Template Function torch_tensorrt::ptq::make_int8_calibrator", "Function torch_tensorrt::torchscript::check_method_operator_support", "Function torch_tensorrt::torchscript::compile", "Function torch_tensorrt::torchscript::embed_engine_in_new_module", "Function torch_tensorrt::get_build_info", "Function torch_tensorrt::set_device", "Function torch_tensorrt::dump_build_info", "Function torch_tensorrt::torchscript::convert_method_to_trt_engine", "Namespace torch_tensorrt", "Namespace torch_tensorrt::logging", "Namespace torch_tensorrt::ptq", "Namespace torch_tensorrt::torchscript", "Program Listing for File logging.h", "Program Listing for File macros.h", "Program Listing for File ptq.h", "Program Listing for File torch_tensorrt.h", "Struct Device", "Struct GraphInputs", "Struct Input", "Struct CompileSpec", "Torch-TensorRT C++ API", "Full API", "torchtrtc", "Conversion Phase", "Writing Dynamo Converters", "Lowering Phase", "Partitioning Phase", "Compiler Phases", "Runtime Phase", "System Overview", "Writing TorchScript Converters", "Useful Links for Torch-TensorRT Development", "Writing Dynamo ATen Lowering Passes", "Compiling Exported Programs with Torch-TensorRT", "TensorRT Backend for torch.compile", "Torch-TensorRT (FX Frontend) User Guide", "Installation", "Overview", "Quick Start", "Torch-TensorRT", "Operators Supported", "torch_tensorrt.dynamo", "torch_tensorrt.fx", "torch_tensorrt.logging", "torch_tensorrt.ts.ptq", "torch_tensorrt.runtime", "torch_tensorrt", "torch_tensorrt.ts", "Computation times", "Changelog", "Configuration", "5. :mod:`test_py_module`", "3. Paragraph Level Markup", "4. Lists & Tables", "1. Long Sticky Nav", "1. Structural Elements", "<no title>", "Installation", "Creating a TorchScript Module", "Using Torch-TensorRT in C++", "Using Torch-TensorRT in Python", "Post Training Quantization (PTQ)", "Using Torch-TensorRT TorchScript Frontend Directly From PyTorch", "Automatically Generate a Converter for a Custom Kernel", "Overloading Torch-TensorRT Converters with Custom Converters", "Cross runtime compilation for windows example", "Using Custom Kernels within TensorRT Engines with Torch-TensorRT", "Engine Caching (BERT)", "Engine Caching", "Dependencies", "Mutable Torch TensorRT Module", "Pre-allocated output buffer", "Refitting Torch-TensorRT Programs with New Weights", "Torch Compile Advanced Usage", "Compiling GPT2 using the Torch-TensorRT torch.compile frontend", "Compiling ResNet with dynamic shapes using the torch.compile backend", "Compiling Stable Diffusion model using the torch.compile backend", "Compiling BERT using the torch.compile backend", "Torch Export with Cudagraphs", "Compiling GPT2 using the dynamo backend", "Compiling Llama2 using the dynamo backend", "Compiling SAM2 using the dynamo backend", "Deploy Quantized Models using Torch-TensorRT", "Weight Streaming", "Torch-TensorRT Tutorials", "Serving a Torch-TensorRT model with Triton", "Legacy notebooks", "Serving a Torch-TensorRT model with Triton", "Dynamic shapes with Torch-TensorRT", "Compile Mixed Precision models with Torch-TensorRT", "Deploying Torch-TensorRT Programs", "Saving models compiled with Torch-TensorRT", "Torch-TensorRT Explained", "DLA"], "terms": {"defin": [0, 1, 2, 3, 4, 16, 17, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 43, 46, 47, 48, 49, 51, 52, 54, 65, 68, 75, 76, 80, 88, 89, 90, 91, 93, 94, 96, 98, 103, 107, 108, 109, 110, 116], "file": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 16, 17, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 46, 47, 48, 49, 52, 54, 56, 58, 59, 64, 65, 66, 67, 68, 71, 72, 74, 76, 77, 78, 80, 81, 83, 87, 89, 91, 95, 114, 115, 117, 118, 121], "torch_tensorrt": [0, 1, 2, 14, 16, 17, 42, 43, 44, 46, 47, 48, 49, 50, 51, 52, 54, 56, 62, 63, 64, 65, 68, 69, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 106, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 123], "h": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 15, 16, 17, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 46, 47, 48, 49, 50, 51, 52, 55, 68, 76, 89, 91, 111], "support": [0, 1, 2, 27, 31, 46, 48, 49, 52, 54, 56, 61, 63, 65, 67, 68, 69, 72, 75, 76, 77, 80, 81, 88, 89, 90, 93, 94, 96, 101, 102, 104, 105, 107, 109, 110, 112, 113, 114, 115, 117, 119, 122, 123], "data": [0, 2, 3, 4, 29, 30, 44, 46, 48, 49, 52, 53, 56, 57, 59, 60, 64, 65, 70, 71, 72, 74, 76, 77, 82, 86, 90, 91, 93, 96, 98, 104, 111, 112, 113, 116], "type": [0, 1, 2, 30, 49, 50, 52, 53, 56, 58, 60, 62, 63, 64, 65, 71, 72, 74, 75, 76, 77, 82, 89, 90, 91, 93, 94, 95, 96, 98, 111, 112, 113, 116, 119, 121], "can": [0, 1, 4, 29, 30, 37, 46, 47, 48, 49, 52, 53, 54, 55, 56, 57, 58, 59, 60, 62, 63, 64, 65, 66, 67, 68, 71, 74, 75, 76, 77, 80, 82, 88, 89, 90, 91, 92, 93, 94, 95, 96, 98, 100, 101, 102, 103, 104, 107, 108, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122], "us": [0, 1, 2, 3, 4, 29, 30, 32, 35, 37, 43, 44, 45, 46, 48, 49, 52, 53, 54, 56, 58, 59, 60, 62, 63, 64, 65, 66, 67, 68, 69, 71, 72, 74, 75, 76, 77, 78, 80, 81, 82, 83, 88, 91, 95, 98, 99, 100, 102, 113, 114, 115, 117, 119, 120, 121, 122, 123], "tensorrt": [0, 1, 3, 4, 29, 30, 31, 32, 33, 36, 37, 44, 45, 46, 48, 49, 52, 53, 54, 55, 56, 57, 59, 60, 62, 67, 68, 71, 72, 74, 75, 76, 77, 88, 91, 95, 98, 99, 101, 103, 105, 106, 107, 108, 113], "engin": [0, 1, 17, 32, 33, 37, 45, 46, 48, 49, 52, 53, 56, 57, 59, 62, 63, 64, 69, 71, 72, 75, 76, 77, 80, 89, 90, 91, 92, 93, 94, 99, 101, 102, 104, 105, 107, 113, 114, 118, 120, 122, 123], "thi": [0, 1, 2, 29, 30, 42, 43, 44, 45, 46, 47, 48, 49, 52, 53, 54, 55, 56, 57, 58, 59, 60, 62, 63, 64, 65, 66, 67, 68, 71, 72, 75, 76, 77, 80, 81, 82, 84, 85, 88, 89, 91, 92, 93, 94, 96, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122], "compat": [0, 1, 46, 55, 58, 64, 65, 71, 75, 76, 77, 111, 122], "c10": [0, 1, 45, 46, 48, 49, 89, 91], "check": [0, 1, 31, 46, 52, 55, 60, 65, 67, 71, 75, 77, 89, 96, 100, 102, 114, 115, 117, 120], "trt": [0, 1, 3, 4, 46, 48, 53, 55, 58, 60, 62, 64, 65, 67, 68, 70, 71, 75, 76, 89, 94, 96, 101, 104, 107, 109, 110, 111, 113, 118, 120, 121], "so": [0, 44, 52, 53, 54, 55, 58, 59, 60, 62, 64, 65, 66, 67, 72, 75, 76, 81, 82, 83, 89, 91, 93, 94, 96, 98, 103, 104, 105, 107, 109, 110, 118], "should": [0, 3, 4, 29, 45, 49, 52, 53, 54, 55, 56, 57, 59, 60, 63, 64, 65, 67, 71, 75, 76, 77, 80, 82, 85, 91, 94, 96, 97, 98, 101, 102, 104, 108, 111, 114, 115, 117], "reason": [0, 65, 88, 94, 96, 98, 122], "you": [0, 1, 2, 29, 30, 46, 48, 49, 52, 53, 54, 55, 56, 58, 59, 60, 63, 65, 66, 67, 68, 71, 75, 76, 77, 80, 82, 83, 84, 88, 89, 90, 91, 92, 93, 94, 96, 98, 99, 100, 102, 108, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122], "need": [0, 1, 2, 25, 29, 43, 46, 53, 54, 55, 60, 65, 66, 67, 71, 72, 75, 76, 82, 89, 90, 91, 93, 94, 96, 97, 98, 100, 102, 114, 115, 116, 117, 118, 120], "explictli": 0, "public": [0, 1, 2, 3, 4, 44, 45, 46, 47, 48, 49, 83, 91], "enum": [0, 1, 2, 42, 45, 46, 51, 71, 77, 91, 94], "valu": [0, 1, 2, 16, 17, 45, 46, 48, 53, 56, 58, 60, 63, 70, 71, 74, 76, 80, 89, 100, 103, 104, 105, 107, 113, 116], "underli": [0, 1, 2, 46, 60], "In": [0, 1, 2, 46, 53, 54, 56, 57, 58, 59, 60, 64, 65, 66, 75, 76, 82, 83, 85, 90, 91, 93, 94, 96, 100, 104, 111, 114, 115, 116, 117, 118, 119, 120, 121], "case": [0, 1, 2, 46, 49, 53, 54, 56, 58, 60, 62, 64, 65, 66, 67, 75, 76, 91, 93, 94, 96, 100, 101, 102, 118, 119, 120], "itself": [0, 1, 2, 46, 52, 55, 92, 94, 114, 115, 117], "interfac": [0, 1, 2, 46, 58, 59, 60, 64, 69, 91], "vs": [0, 1, 2, 46, 55, 66, 71, 76, 77, 92], "normal": [0, 1, 2, 46, 65, 82, 88, 89, 91, 94, 100, 101, 102, 108, 111, 112, 114, 115, 117, 123], "instatin": [0, 1, 2, 46], "ex": [0, 1, 2, 33, 46, 67, 77, 83, 85], "kfloat": [0, 45, 49], "enumer": [0, 1, 2, 16, 17, 46, 111], "klong": [0, 45], "int64": [0, 76, 77, 113], "kdoubl": [0, 45], "fp64": [0, 76], "fp32": [0, 48, 49, 52, 64, 65, 71, 76, 77, 91, 109, 110, 111, 114, 115, 116, 117, 119], "khalf": [0, 45, 89], "fp16": [0, 48, 49, 52, 64, 65, 71, 72, 76, 89, 90, 100, 106, 109, 110, 111, 113, 119, 123], "kchar": [0, 45], "int8": [0, 44, 48, 49, 52, 64, 71, 76, 77, 91, 112, 123], "kint": [0, 45], "int": [0, 3, 4, 35, 44, 45, 49, 52, 54, 56, 63, 64, 70, 71, 72, 76, 77, 80, 89, 93, 96, 111, 112, 113], "kbool": [0, 45], "bool": [0, 1, 2, 3, 4, 24, 27, 30, 31, 42, 44, 45, 46, 49, 55, 60, 64, 70, 71, 72, 74, 75, 76, 77, 80, 89, 91, 95, 96], "kunknown": [0, 2, 45], "sentinel": [0, 2, 76], "function": [0, 1, 2, 3, 4, 46, 48, 49, 51, 54, 55, 56, 58, 60, 62, 64, 65, 66, 88, 89, 91, 92, 93, 94, 96, 102, 103, 107, 108, 109, 110, 111, 114, 115, 116, 117, 118, 120, 122, 123], "default": [0, 1, 2, 3, 4, 16, 29, 30, 33, 43, 45, 46, 48, 49, 52, 54, 56, 62, 64, 65, 66, 71, 72, 75, 76, 77, 80, 81, 82, 89, 90, 91, 92, 93, 94, 95, 96, 98, 112, 118, 120, 121, 122], "construct": [0, 1, 2, 3, 4, 46, 48, 49, 53, 54, 55, 57, 59, 60, 65, 74, 75, 76, 82, 83, 89, 91, 94, 96, 98, 118], "new": [0, 1, 2, 3, 4, 32, 33, 46, 48, 49, 56, 58, 59, 60, 62, 64, 65, 68, 69, 71, 77, 82, 89, 93, 98, 99, 100, 101, 104, 105, 107, 108, 114, 115, 117, 120], "object": [0, 1, 2, 3, 4, 46, 48, 49, 52, 58, 60, 62, 63, 64, 71, 75, 76, 77, 91, 92, 94, 101, 118, 121], "inlin": [0, 1, 2, 3, 4, 29, 30, 44, 46, 48, 55, 83, 86, 89], "constexpr": [0, 1, 2, 45, 46, 93, 96], "t": [0, 1, 2, 45, 46, 55, 60, 65, 66, 70, 76, 80, 82, 83, 88, 89, 91, 93, 94, 96, 112, 114, 115, 117, 118], "constructor": [0, 2, 46, 48, 49, 58, 88], "from": [0, 1, 2, 3, 4, 29, 30, 44, 46, 48, 49, 52, 53, 55, 56, 57, 58, 59, 60, 63, 64, 65, 67, 69, 71, 72, 75, 76, 77, 78, 80, 81, 82, 83, 88, 89, 91, 93, 94, 95, 96, 97, 98, 100, 101, 102, 104, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 120, 121, 122], "torchtrt_api": [0, 2, 19, 22, 23, 24, 25, 26, 27, 28, 31, 32, 33, 34, 35, 36, 37, 42, 43, 44, 45, 48, 49, 50], "scalartyp": [0, 45, 70], "torch": [0, 1, 2, 4, 20, 21, 29, 30, 31, 32, 33, 36, 37, 44, 45, 46, 47, 48, 49, 52, 53, 54, 55, 56, 57, 58, 59, 60, 62, 67, 71, 72, 74, 75, 76, 77, 78, 88, 91, 93, 95, 97, 98, 99, 101, 113, 123], "paramet": [0, 1, 2, 3, 4, 25, 26, 27, 29, 30, 31, 32, 33, 35, 37, 46, 48, 49, 53, 54, 55, 60, 64, 65, 71, 72, 74, 75, 76, 77, 86, 88, 89, 102, 109, 110], "oper": [0, 1, 2, 3, 4, 31, 44, 45, 46, 49, 52, 53, 55, 56, 57, 58, 59, 60, 62, 63, 65, 69, 71, 76, 77, 90, 91, 94, 101, 102, 105, 107, 108, 111, 122, 123], "const": [0, 1, 2, 3, 4, 29, 30, 31, 32, 33, 35, 37, 44, 45, 46, 55, 60, 70, 89, 91], "get": [0, 1, 2, 3, 4, 23, 34, 44, 46, 55, 56, 60, 62, 63, 65, 67, 75, 76, 89, 91, 93, 94, 98, 104, 109, 110, 113, 114, 115, 116, 117], "return": [0, 1, 2, 3, 4, 23, 24, 29, 30, 31, 32, 33, 34, 37, 42, 43, 44, 45, 46, 54, 55, 56, 57, 58, 59, 60, 62, 64, 65, 71, 72, 75, 76, 77, 88, 89, 90, 91, 93, 94, 96, 98, 101, 102, 103, 108, 111, 112, 113, 114, 115, 117, 118, 119], "explicit": [0, 1, 2, 3, 4, 45, 46, 55, 65, 72, 75, 82, 91, 122], "delet": [0, 1, 2, 45, 46, 55], "other": [0, 1, 2, 45, 46, 52, 53, 55, 58, 62, 64, 65, 66, 70, 71, 75, 76, 81, 82, 89, 90, 94, 120], "comparis": [0, 2], "true": [0, 1, 2, 4, 46, 49, 55, 56, 60, 62, 64, 65, 70, 71, 72, 75, 76, 77, 80, 83, 89, 91, 92, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 117, 119, 120, 123], "fals": [0, 1, 2, 3, 4, 44, 45, 46, 49, 54, 62, 64, 65, 70, 71, 72, 75, 76, 77, 80, 81, 82, 83, 89, 91, 92, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 106, 107, 108, 109, 110, 111, 112, 113, 120], "struct": [1, 21, 38, 41, 45, 54, 91], "onli": [1, 3, 4, 16, 29, 44, 46, 48, 52, 54, 55, 56, 59, 60, 64, 65, 67, 68, 71, 72, 75, 76, 82, 91, 93, 94, 95, 96, 100, 102, 110, 113, 119, 120, 123], "applic": [1, 29, 46, 52, 55, 59, 64, 71, 75, 76, 89, 90, 92, 120, 123], "kcuda": [1, 46, 56, 89], "which": [1, 2, 29, 32, 37, 46, 49, 53, 54, 55, 56, 57, 58, 59, 60, 62, 63, 64, 65, 66, 71, 72, 74, 75, 76, 77, 80, 82, 83, 88, 89, 90, 91, 92, 93, 94, 95, 96, 98, 103, 104, 105, 108, 109, 110, 111, 114, 115, 116, 117, 118, 119, 120, 121, 122], "map": [1, 46, 53, 54, 55, 57, 59, 60, 65, 75, 76, 89, 91, 92, 98, 103, 114, 115, 116, 117], "kgpu": [1, 45, 46], "To": [1, 46, 52, 54, 56, 64, 66, 71, 80, 88, 89, 90, 92, 96, 102, 109, 110, 111, 114, 115, 117], "datatyp": [1, 21, 38, 45, 46, 48, 49, 50, 71, 76, 77, 90, 96, 114, 115, 117, 119], "target": [1, 33, 45, 46, 48, 49, 52, 54, 56, 58, 59, 64, 65, 66, 69, 71, 75, 76, 77, 90, 91, 92, 94, 96, 102, 122, 123], "gpu": [1, 32, 35, 37, 45, 46, 52, 64, 65, 71, 75, 76, 77, 89, 91, 92, 93, 96, 101, 104, 109, 110, 113, 114, 115, 117, 120, 122, 123], "run": [1, 37, 46, 49, 52, 53, 54, 55, 56, 57, 58, 59, 60, 64, 65, 66, 67, 68, 71, 72, 75, 76, 77, 82, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123], "kdla": [1, 45, 46, 123], "dla": [1, 45, 46, 49, 52, 64, 69, 71, 76, 77], "intern": [1, 16, 46, 60, 63, 73, 75, 82, 89], "note": [1, 46, 48, 54, 60, 62, 65, 66, 67, 75, 76, 80, 82, 89, 96, 102, 108, 114, 115, 117, 118, 123], "The": [1, 46, 48, 49, 52, 53, 54, 55, 56, 57, 58, 59, 60, 62, 63, 64, 65, 66, 71, 75, 76, 77, 80, 83, 88, 90, 91, 92, 93, 94, 95, 96, 97, 98, 100, 101, 102, 104, 105, 108, 109, 111, 113, 114, 115, 116, 117, 118, 121, 122], "valid": [1, 46, 56, 60, 62, 71, 75, 76, 94], "kcpu": [1, 46], "comparison": [1, 46], "an": [2, 3, 4, 48, 49, 52, 53, 54, 55, 56, 57, 58, 59, 60, 62, 64, 65, 66, 68, 71, 72, 74, 75, 76, 77, 80, 82, 83, 88, 89, 90, 91, 93, 94, 96, 98, 102, 103, 104, 108, 109, 110, 113, 114, 115, 116, 117, 118, 120, 121, 122], "memeori": 2, "layout": [2, 48, 70, 71, 76, 77], "store": [2, 4, 49, 52, 53, 58, 60, 64, 65, 71, 75, 76, 77, 88, 89, 93, 96, 98, 102, 111], "tensor": [2, 33, 44, 45, 48, 49, 52, 53, 54, 55, 56, 58, 60, 62, 63, 64, 65, 70, 71, 72, 75, 76, 77, 88, 89, 90, 91, 93, 94, 96, 101, 103, 108, 111, 113, 116], "kcontigu": [2, 45, 48], "contigu": [2, 48, 49, 52, 71, 76, 77], "nchw": [2, 71, 76, 77], "linear": [2, 56, 70, 76, 88, 96, 112, 119], "kchannelslast": [2, 45], "channel": [2, 76, 81], "last": [2, 55, 65, 76, 112], "nhwc": [2, 52], "memoryformat": [2, 45], "ptq": [3, 4, 15, 18, 19, 38, 50, 51, 52, 69, 71, 76, 77], "privat": [3, 4, 44, 45, 91], "algorithm": [3, 4, 29, 30, 44, 65, 74, 91, 110], "typenam": [3, 4, 29, 30, 44], "gener": [3, 4, 29, 52, 55, 58, 59, 60, 62, 64, 65, 66, 71, 72, 80, 82, 83, 86, 88, 89, 91, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 111, 112, 113, 114, 120], "int8calibr": [3, 20, 30, 40, 44, 50], "implement": [3, 4, 55, 56, 58, 63, 65, 75, 81, 89, 91, 93, 96, 98, 111, 120], "specifi": [3, 4, 33, 52, 54, 60, 64, 65, 66, 71, 76, 77, 80, 82, 90, 92, 113, 114, 115, 117, 118, 119, 121, 122], "calibr": [3, 4, 29, 30, 44, 49, 52, 71, 74, 76, 77, 89, 91], "read": [3, 4, 29, 30, 44, 80, 82, 91, 111], "nvinfer1": [3, 4, 29, 30, 44, 45, 49, 60, 91], "iint8calibr": [3, 4, 29, 30, 44, 45, 49, 71, 76, 77, 91], "iint8entropycalibrator2": [3, 4, 29, 30, 44, 91], "std": [3, 4, 22, 26, 28, 29, 30, 31, 33, 34, 37, 42, 44, 45, 47, 48, 49, 56, 89, 91, 114, 115, 117, 123], "string": [3, 4, 18, 20, 21, 22, 26, 28, 29, 30, 31, 33, 34, 37, 42, 44, 45, 49, 54, 56, 58, 60, 64, 71, 76, 80, 89, 91], "cache_file_path": [3, 4, 29, 30, 44], "8": [3, 52, 55, 63, 64, 66, 75, 76, 82, 83, 86, 89, 95, 96, 105, 108, 114, 115, 117, 118], "cach": [3, 4, 29, 30, 44, 52, 64, 65, 69, 71, 72, 74, 76, 89, 91, 99, 101, 114, 120], "getbatchs": [3, 4, 44], "noexcept": [3, 4, 44, 91], "overrid": [3, 4, 29, 30, 44, 54, 65, 91], "batch": [3, 4, 44, 64, 65, 72, 75, 91, 98, 105, 107, 112, 113, 114, 115, 117, 118, 123], "size": [3, 4, 44, 48, 49, 52, 55, 56, 64, 65, 70, 71, 72, 76, 77, 80, 89, 91, 93, 96, 98, 105, 107, 111, 112, 116, 118], "next": [3, 4, 53, 54, 58, 63, 72, 76, 80, 82, 83, 91, 94, 101, 103, 108, 112, 114, 115, 117], "alwai": [3, 4, 27, 52, 76, 82, 102, 113], "1": [3, 4, 33, 44, 45, 48, 49, 52, 54, 55, 56, 58, 60, 62, 63, 64, 65, 66, 70, 71, 72, 74, 75, 76, 77, 79, 80, 82, 83, 86, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 100, 101, 102, 104, 105, 107, 108, 109, 110, 111, 112, 113, 116, 118, 119, 121, 123], "due": [3, 4, 66, 81, 82, 104, 112], "issu": [3, 4, 64, 71, 76, 89, 103, 104, 107], "getbatch": [3, 4, 44], "void": [3, 4, 25, 26, 27, 28, 35, 36, 42, 44, 45], "bind": [3, 4, 33, 44, 75, 77, 82], "char": [3, 4, 44, 52, 89], "name": [3, 4, 31, 33, 37, 44, 54, 56, 58, 60, 65, 66, 67, 72, 74, 75, 76, 77, 82, 83, 88, 89, 92, 93, 94, 96, 102, 108, 112, 114, 115, 117, 119], "nbbind": [3, 4, 44], "Not": 3, "arrai": [3, 4, 33, 53, 54, 76, 77, 94, 96, 101, 111, 113], "pointer": [3, 4, 91], "fed": [3, 4, 48], "buffer": [3, 4, 65, 69, 96, 99, 114], "each": [3, 4, 49, 53, 55, 56, 58, 60, 64, 65, 66, 71, 72, 75, 80, 82, 89, 93, 94, 102, 108, 110, 120], "input": [3, 4, 21, 29, 33, 38, 44, 45, 47, 49, 50, 52, 53, 54, 55, 56, 58, 60, 62, 63, 64, 65, 68, 70, 71, 72, 73, 75, 76, 77, 83, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 108, 109, 110, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123], "number": [3, 4, 49, 52, 54, 55, 56, 60, 63, 64, 65, 71, 72, 76, 77, 80, 89, 90, 96, 102, 104, 105, 107, 113, 116, 122], "readcalibrationcach": [3, 4, 44], "size_t": [3, 4, 44, 91], "length": [3, 4, 44, 65, 70, 83, 104, 113], "how": [3, 4, 66, 67, 82, 84, 86, 88, 92, 93, 94, 96, 98, 100, 103, 112, 113, 114, 115, 116, 117, 118, 120], "enabl": [3, 4, 24, 49, 52, 54, 56, 57, 59, 64, 65, 66, 71, 72, 74, 75, 76, 77, 80, 98, 100, 102, 105, 107, 108, 109, 110, 111, 113, 119, 120], "use_cach": [3, 4, 30, 44, 74, 91, 104, 109, 110, 113], "set": [3, 4, 16, 21, 25, 27, 29, 32, 35, 37, 45, 46, 48, 49, 52, 53, 54, 55, 56, 57, 58, 59, 65, 66, 71, 72, 75, 76, 77, 80, 84, 87, 88, 89, 90, 91, 93, 94, 96, 101, 102, 109, 112, 113, 116, 118, 119, 120, 122, 123], "writecalibrationcach": [3, 4, 44], "write": [3, 4, 29, 30, 44, 65, 69, 82, 89, 91, 114, 115, 117], "provid": [3, 4, 49, 52, 54, 56, 58, 60, 62, 64, 65, 66, 68, 71, 72, 75, 76, 77, 82, 89, 90, 91, 92, 93, 94, 98, 99, 102, 103, 104, 108, 111, 113, 114, 115, 117, 118, 120, 121, 122], "cast": [3, 4, 55, 64, 71, 109, 110, 111, 119], "convienc": [3, 4, 49], "convert": [3, 4, 31, 32, 37, 52, 55, 56, 57, 59, 63, 64, 69, 71, 76, 77, 90, 92, 96, 99, 104, 105, 107, 111, 113, 114, 116, 120], "easili": [3, 4, 100], "assign": [3, 4, 81], "ptq_calibr": [3, 4, 45, 49, 91], "field": [3, 4, 63, 72, 76, 91], "compilespec": [3, 4, 21, 32, 37, 41, 45, 50, 56, 77, 89, 91, 123], "dataloaderuniqueptr": [4, 44], "libtorch": [4, 36, 60, 66, 68, 89, 91, 122], "dataload": [4, 29, 30, 44, 49, 74, 91, 112], "unique_ptr": [4, 30], "unqiue_ptr": 4, "A": [4, 29, 30, 32, 33, 47, 48, 54, 55, 56, 60, 65, 66, 71, 72, 76, 77, 83, 91, 93, 106, 114, 115, 117], "uniqu": [4, 90], "what": [4, 54, 55, 65, 68, 76, 82, 88, 89, 90, 104, 109, 110, 122], "make_data_load": [4, 91], "factori": [4, 29, 30, 64, 71, 91], "path": [4, 13, 14, 15, 29, 30, 52, 64, 65, 66, 67, 71, 74, 76, 88, 89, 91, 95, 98, 108, 112, 122], "find": [4, 65, 66, 67, 89, 96, 113], "whether": [4, 52, 54, 64, 65, 71, 72, 76, 81, 91, 105, 107, 120], "exist": [4, 31, 32, 37, 54, 63, 64, 65, 67, 71, 74, 76, 77, 91, 98, 116], "There": [4, 53, 54, 59, 60, 62, 63, 65, 66, 83, 88, 91, 102, 114, 115, 116, 117, 118, 120], "consum": [4, 53, 88], "macro": [5, 6, 7, 8, 9, 10, 11, 12, 15, 18, 20, 21, 42, 44, 45, 50, 51], "x": [5, 10, 33, 43, 55, 56, 66, 67, 68, 75, 77, 83, 88, 89, 93, 94, 96, 98, 103, 108, 112, 113, 114, 115, 117, 118, 119, 121], "includ": [13, 15, 16, 34, 36, 42, 43, 44, 45, 51, 52, 54, 56, 57, 58, 59, 62, 64, 65, 66, 67, 68, 71, 72, 75, 76, 80, 82, 88, 89, 91, 96, 111, 120], "parent": [14, 15, 18, 19, 20, 21], "cpp": [14, 15, 42, 43, 44, 45, 51, 55, 59, 66, 89, 91], "log": [15, 16, 19, 20, 38, 44, 50, 51, 55, 60, 64, 65, 69, 70, 71, 72, 76, 93, 94, 105, 107, 119], "emum": [16, 17], "messag": [16, 25, 26, 52, 73], "sever": [16, 26, 73, 108], "kinternal_error": [16, 42], "print": [16, 31, 44, 62, 64, 67, 71, 77, 82, 89, 92, 93, 94, 95, 96, 97, 98, 100, 101, 102, 104, 105, 107, 109, 110, 112, 113, 114, 115, 117], "error": [16, 49, 52, 53, 55, 59, 64, 65, 71, 73, 76, 77, 82, 89, 93, 118], "kerror": [16, 42], "all": [16, 42, 43, 44, 45, 49, 52, 54, 55, 56, 58, 62, 64, 65, 66, 67, 71, 73, 75, 76, 78, 82, 83, 88, 89, 90, 91, 94, 96, 109, 110, 114, 115, 116, 117, 119, 120, 122], "kwarn": [16, 42], "warn": [16, 44, 52, 60, 73, 75], "kinfo": [16, 42, 44], "info": [16, 32, 37, 45, 52, 60, 73, 75, 76, 119], "kdebug": [16, 42, 44], "debug": [16, 27, 45, 49, 52, 60, 62, 64, 71, 73, 75, 76, 77, 92, 93, 95, 96, 97, 98, 100, 102, 103, 105, 107, 112, 119], "kgraph": [16, 42, 55], "everyth": [16, 64, 71, 76], "intermedi": [16, 49, 52, 54, 64, 71, 73, 76, 77, 88, 119, 122], "graph": [16, 31, 32, 37, 45, 49, 52, 53, 54, 56, 57, 59, 60, 62, 63, 64, 65, 71, 72, 73, 76, 77, 88, 89, 93, 94, 96, 98, 100, 101, 102, 104, 105, 107, 111, 116, 118, 120], "lower": [16, 54, 63, 65, 69, 71, 72, 73, 76, 83, 93, 96, 98, 105, 107, 113, 116, 122], "phase": [16, 60, 63, 89, 94, 101, 102, 108, 118, 122], "class": [17, 29, 30, 44, 45, 46, 51, 58, 60, 64, 65, 73, 77, 82, 83, 88, 89, 90, 91, 93, 94, 96, 98, 103, 104, 108, 111, 112, 116, 118, 119], "int8_t": [17, 45], "select": [17, 29, 30, 37, 49, 52, 58, 64, 65, 66, 70, 71, 76, 77, 81, 84, 90, 91, 96, 111, 122], "capabl": [17, 45, 49, 52, 58, 71, 76, 77, 92, 94, 95], "kstandard": [17, 45, 49], "ksafeti": [17, 45], "kdla_standalon": [17, 45], "directori": [18, 19, 20, 21, 42, 43, 44, 45, 50, 66, 67, 71, 91, 98, 111, 114, 115, 117], "program": [18, 19, 20, 21, 29, 51, 52, 57, 58, 59, 69, 71, 88, 93, 98, 99, 109, 110, 114, 118], "list": [18, 19, 20, 21, 31, 49, 51, 53, 56, 58, 60, 62, 63, 65, 68, 70, 71, 72, 75, 76, 77, 86, 89, 90, 94, 96, 114, 115, 117], "level": [18, 23, 25, 26, 39, 42, 44, 50, 54, 55, 56, 59, 64, 65, 71, 76, 77, 86, 88, 94, 96, 114, 115, 117, 122], "get_is_colored_output_on": [18, 39, 42, 50], "get_logging_prefix": [18, 39, 42, 50], "get_reportable_log_level": [18, 39, 42, 50], "set_is_colored_output_on": [18, 39, 42, 50], "set_logging_prefix": [18, 39, 42, 50], "set_reportable_log_level": [18, 39, 42, 50], "torchscript": [19, 21, 38, 43, 45, 49, 50, 52, 56, 57, 58, 59, 63, 68, 71, 72, 74, 75, 76, 77, 90, 101, 114, 115, 116, 117, 118, 123], "str": [19, 43, 44, 50, 54, 64, 65, 70, 71, 74, 75, 76, 77, 94, 95, 96, 98, 112], "torch_tensorrt_major_vers": [19, 43, 50], "torch_tensorrt_minor_vers": [19, 43, 50], "torch_tensorrt_patch_vers": [19, 43, 50], "torch_tensorrt_vers": [19, 43, 50], "torchtrt_hidden": [19, 43, 50], "xstr": [19, 43, 50], "nvinfer": [20, 44], "fstream": [20, 44], "iostream": [20, 21, 44, 45, 89], "iter": [20, 44, 49, 52, 53, 64, 71, 74, 76, 77, 97, 98, 111, 112, 113], "memori": [20, 21, 44, 45, 55, 60, 71, 76, 77, 89, 90, 93, 96, 98, 101, 108, 109, 110, 113], "sstream": [20, 44], "vector": [20, 21, 33, 44, 45, 47, 48, 49, 56, 58, 76, 89, 91, 123], "templat": [20, 40, 44, 45, 50, 80, 89], "int8cachecalibr": [20, 29, 40, 44, 50], "make_int8_cache_calibr": [20, 40, 44, 50, 91], "make_int8_calibr": [20, 29, 40, 44, 50, 91], "cuda_runtim": [21, 45], "custom_class": [21, 45], "devic": [21, 33, 35, 38, 45, 49, 50, 52, 58, 64, 70, 71, 72, 74, 75, 76, 77, 90, 91, 92, 93, 96, 100, 104, 106, 109, 110, 111, 113, 116, 123], "graphinput": [21, 38, 45, 49, 50], "devicetyp": [21, 38, 45, 46, 50, 75, 76, 77, 91, 92, 96, 123], "tensorformat": [21, 38, 45, 48, 50, 76, 96], "enginecap": [21, 38, 45, 49, 50, 64, 71, 75, 76, 77, 92, 96], "dump_build_info": [21, 38, 45, 50], "get_build_info": [21, 38, 45, 50], "set_devic": [21, 38, 45, 50, 120], "check_method_operator_support": [21, 41, 45, 50], "compil": [21, 31, 37, 41, 45, 49, 50, 52, 54, 55, 56, 58, 60, 62, 65, 71, 72, 73, 75, 76, 77, 78, 80, 88, 90, 91, 92, 93, 94, 96, 97, 99, 100, 112, 114, 115, 117, 120, 123], "convert_method_to_trt_engin": [21, 41, 45, 50, 76, 77, 89, 92], "embed_engine_in_new_modul": [21, 41, 45, 50, 77], "current": [23, 54, 56, 58, 60, 62, 63, 64, 65, 66, 67, 71, 72, 76, 77, 80, 94, 96, 100, 104, 109, 110, 111, 112, 113, 120], "report": [23, 44, 75], "Is": [24, 76], "color": [24, 27, 82, 111], "output": [24, 27, 33, 49, 52, 53, 54, 55, 56, 58, 60, 62, 63, 64, 65, 66, 69, 71, 73, 75, 76, 77, 80, 82, 83, 89, 93, 94, 96, 98, 99, 100, 102, 106, 113, 114, 115, 116, 117, 118, 119, 121], "lvl": [25, 26, 42], "inform": [25, 33, 34, 36, 48, 52, 53, 56, 58, 62, 64, 65, 66, 71, 72, 73, 76, 82, 88, 89, 91, 92, 96, 98, 113, 118], "ad": [25, 52, 53, 54, 56, 62, 65, 66, 93, 96, 100], "abov": [25, 54, 56, 62, 65, 66, 73, 81, 82, 89, 96, 105, 107, 111, 119, 121], "msg": [26, 42], "add": [26, 53, 54, 55, 56, 60, 63, 66, 70, 80, 82, 87, 89, 90, 93, 94, 96], "global": [26, 52, 64, 71, 76, 89], "colored_output_on": [27, 42], "prefix": [27, 28, 42, 82], "help": [27, 52, 53, 60, 64, 65, 89, 95, 98, 108, 112, 113, 116, 120], "when": [27, 44, 45, 46, 52, 53, 55, 56, 57, 58, 59, 60, 64, 65, 66, 71, 75, 76, 77, 80, 82, 84, 88, 89, 91, 94, 96, 98, 100, 101, 102, 108, 113, 116, 118, 120], "termin": [27, 52, 89], "If": [27, 33, 53, 54, 55, 56, 62, 63, 64, 65, 66, 68, 71, 72, 76, 80, 82, 89, 90, 91, 94, 96, 98, 101, 102, 103, 108, 113, 114, 115, 117, 118, 119, 120, 122, 123], "build": [29, 30, 34, 49, 52, 53, 57, 59, 60, 63, 64, 65, 71, 75, 76, 81, 86, 89, 91, 93, 94, 96, 105, 107, 113, 118], "post": [29, 30, 49, 52, 63, 69, 89, 98], "train": [29, 30, 49, 52, 69, 70, 89, 90, 98, 113], "quantiz": [29, 30, 52, 64, 69, 74, 76, 89, 99, 114], "creat": [29, 30, 33, 52, 53, 54, 56, 58, 60, 65, 69, 76, 77, 82, 89, 93, 94, 96, 102, 111, 113, 114, 115, 117], "previous": [29, 33, 89, 93, 98, 102], "therefor": [29, 58, 65, 66, 75, 82, 89, 116, 120], "have": [29, 33, 44, 52, 53, 54, 55, 56, 60, 62, 63, 64, 65, 66, 67, 71, 72, 74, 75, 76, 77, 82, 88, 89, 90, 91, 93, 96, 99, 104, 105, 107, 111, 112, 114, 115, 116, 117, 118], "requir": [29, 49, 52, 53, 54, 55, 63, 64, 65, 66, 67, 71, 76, 77, 80, 89, 91, 94, 95, 96, 99, 101, 104, 108, 111, 112, 113, 114, 115, 117, 118, 120], "dataset": [29, 74, 91, 116], "save": [29, 44, 52, 58, 64, 65, 68, 69, 71, 75, 76, 77, 89, 90, 95, 97, 98, 101, 102, 106, 111, 113, 114, 115, 116, 117, 120, 122], "later": [29, 71, 89, 93, 102, 121, 122], "differ": [29, 55, 56, 59, 64, 65, 66, 71, 76, 80, 88, 94, 96, 98, 100, 109, 113, 116, 120, 122], "scratch": [29, 98, 102], "depend": [29, 34, 53, 59, 64, 65, 67, 68, 71, 89, 90, 104, 111, 113, 115, 117, 120], "howev": [29, 66, 80, 81, 89, 93, 94, 96, 98, 114, 115, 117, 118, 122], "network": [29, 30, 54, 60, 65, 76, 89, 91, 94, 96, 113, 114, 115, 116, 117, 123], "also": [29, 53, 54, 60, 62, 64, 66, 68, 80, 82, 83, 89, 90, 91, 93, 98, 108, 111, 112, 116], "recalibr": 29, "its": [29, 53, 56, 58, 60, 66, 75, 76, 82, 93, 96, 112, 114, 115, 117, 120, 122], "structur": [29, 46, 49, 56, 59, 60, 64, 71, 76, 80, 82, 86, 88, 96, 114, 115, 117], "chang": [29, 55, 56, 59, 62, 64, 65, 75, 76, 77, 80, 91, 93, 94, 98, 100, 101, 102, 111, 114, 115, 117, 120, 122], "respons": [29, 54, 58, 82, 120], "ensur": [29, 54, 55, 56, 62, 64, 66, 67, 71, 75, 93, 101, 108, 109, 110, 111], "By": [29, 30, 51, 56, 64, 66, 71, 80, 88, 98, 118], "entropi": [29, 30, 91], "v2": [29, 30, 82], "perform": [29, 30, 54, 62, 63, 71, 75, 76, 91, 93, 96, 108, 111, 113, 114, 115, 116, 117, 119, 120, 121, 122], "recommend": [29, 30, 65, 66, 76, 82, 89, 96, 114, 115, 117, 118], "feed": [29, 30, 89], "forward": [29, 30, 32, 33, 56, 58, 60, 64, 68, 71, 75, 76, 77, 88, 89, 90, 91, 92, 93, 94, 96, 103, 104, 108, 111, 112, 118, 119], "minmax": [29, 30, 91], "recomend": [29, 30], "nlp": [29, 30, 91], "task": [29, 30, 65, 91, 101, 116], "call": [29, 30, 32, 49, 54, 55, 58, 60, 65, 71, 72, 75, 76, 77, 82, 88, 89, 92, 94, 96, 98, 100, 103, 107, 116, 118, 120, 122], "e": [29, 30, 52, 55, 60, 65, 66, 67, 68, 72, 76, 88, 89, 91, 96, 98, 102, 114, 115, 117], "g": [29, 30, 52, 55, 65, 66, 67, 72, 76, 82, 91, 96, 102, 114, 115, 117], "iint8minmaxcalibr": [29, 30, 91], "calibration_cache_fil": [29, 30, 91], "move": [30, 44, 55, 58, 77, 89, 91, 94, 101, 109, 110], "calibration_dataload": [30, 91], "contain": [30, 31, 52, 53, 54, 55, 56, 60, 65, 66, 72, 75, 76, 82, 83, 88, 89, 91, 96, 98, 101, 111, 114, 115, 117, 120], "jit": [31, 32, 33, 37, 45, 47, 49, 52, 53, 55, 56, 57, 58, 59, 60, 61, 64, 68, 69, 71, 75, 76, 77, 88, 89, 90, 92, 93, 96, 102, 114, 115, 117, 121, 122], "modul": [31, 32, 33, 37, 45, 49, 52, 56, 57, 58, 59, 60, 64, 65, 66, 67, 68, 69, 71, 72, 74, 75, 76, 77, 81, 82, 83, 90, 91, 92, 93, 94, 95, 96, 99, 101, 102, 103, 104, 111, 112, 114, 116, 118, 119, 121, 123], "method_nam": [31, 37, 45, 52, 76, 77, 89], "see": [31, 55, 56, 58, 62, 64, 65, 66, 76, 77, 82, 88, 89, 90, 93, 94, 96, 98, 102, 103], "fulli": [31, 52, 55, 64, 71, 75, 76, 77, 89, 91, 96, 123], "take": [31, 32, 33, 37, 53, 54, 57, 58, 59, 60, 62, 65, 71, 72, 75, 76, 77, 80, 82, 89, 91, 92, 94, 96, 103, 116, 118], "method": [31, 32, 33, 37, 48, 52, 55, 60, 66, 71, 76, 77, 82, 88, 89, 92, 98, 116], "pure": [31, 71, 76], "Will": 31, "out": [31, 44, 53, 55, 56, 57, 59, 60, 64, 66, 71, 76, 77, 82, 89, 96, 100, 111, 112, 113, 114, 115, 117, 118], "unsupport": [31, 49, 54, 64, 76, 96, 101, 122], "script": [31, 55, 56, 68, 76, 77, 88, 89, 90, 92, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 117, 120, 122], "nvidia": [32, 37, 42, 43, 44, 45, 52, 61, 64, 65, 66, 67, 71, 76, 77, 89, 103, 107, 114, 115, 117, 122, 123], "configur": [32, 37, 48, 62, 64, 66, 71, 75, 76, 77, 86, 89, 91, 96, 113, 114, 115, 117, 118], "equival": [32, 57, 59, 60, 71, 76, 77, 88, 89, 91, 94, 96, 105, 107], "specif": [32, 49, 54, 55, 57, 59, 62, 64, 71, 76, 77, 82, 94, 113, 122], "traget": 32, "input_binding_nam": [33, 45, 75, 77], "output_binding_nam": [33, 45, 75, 77], "emb": [33, 52, 63, 77, 83], "pre": [33, 55, 69, 74, 77, 91, 98, 99, 113, 114, 120], "built": [33, 52, 58, 59, 64, 66, 71, 75, 76, 77, 98, 102, 111], "serial": [33, 37, 52, 57, 59, 66, 71, 75, 76, 77, 89, 96, 98, 114, 115, 117, 122], "regist": [33, 54, 58, 60, 65, 75, 77, 93, 94, 96], "execut": [33, 49, 52, 55, 57, 58, 59, 63, 64, 65, 66, 69, 71, 72, 75, 76, 77, 78, 88, 89, 91, 94, 96, 101, 108, 114, 115, 117], "must": [33, 48, 49, 52, 54, 55, 56, 60, 62, 65, 66, 71, 72, 76, 77, 82, 83, 89, 93, 98, 118, 120, 122], "follow": [33, 52, 54, 56, 58, 62, 63, 64, 65, 66, 77, 80, 82, 83, 87, 88, 89, 91, 93, 94, 96, 98, 99, 104, 105, 109, 110, 114, 115, 116, 117, 118, 119, 120], "format": [33, 45, 48, 49, 52, 70, 71, 76, 77, 82, 83, 90, 96, 98, 112, 114, 115, 116, 117, 119, 121], "symbol": [33, 65, 66, 77, 82, 120], "index": [33, 61, 62, 66, 67, 69, 70, 77, 80, 86, 91, 96, 111], "0": [33, 43, 44, 45, 49, 52, 54, 56, 59, 60, 62, 64, 65, 66, 67, 69, 70, 71, 72, 74, 75, 76, 77, 78, 79, 81, 82, 89, 91, 92, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 117, 118, 119, 123], "2": [33, 43, 54, 56, 60, 63, 64, 65, 66, 67, 69, 70, 71, 74, 75, 76, 77, 80, 82, 83, 86, 88, 89, 91, 93, 94, 96, 97, 98, 100, 102, 103, 104, 105, 107, 108, 109, 110, 111, 112, 113, 118, 121], "y": [33, 56, 77, 83, 93, 94, 96, 103], "compilesepc": 33, "order": [33, 49, 54, 56, 60, 62, 65, 66, 71, 72, 75, 76, 77, 89, 90, 94, 98, 119], "pass": [33, 53, 54, 56, 57, 58, 59, 60, 63, 64, 65, 66, 69, 73, 74, 75, 76, 77, 88, 89, 91, 93, 94, 96, 98, 102], "origin": [33, 65, 72, 76, 96, 98, 100, 111, 122], "pytorch": [33, 48, 49, 52, 54, 55, 56, 57, 58, 59, 60, 63, 64, 66, 67, 68, 71, 74, 75, 76, 77, 88, 89, 90, 91, 94, 98, 100, 101, 102, 111, 112, 114, 115, 117, 118, 119, 120, 121, 122], "assum": [33, 75, 92, 96, 99, 114], "convent": 33, "below": [33, 56, 60, 62, 63, 64, 65, 66, 67, 82, 89, 90, 98, 101, 106, 111, 114, 115, 117], "librari": [34, 42, 43, 44, 45, 52, 54, 57, 58, 59, 60, 76, 89, 93, 96, 99, 114], "version": [34, 36, 59, 62, 64, 65, 67, 71, 76, 80, 83, 96, 114, 115, 116, 117, 121], "gpu_id": [35, 45, 46, 52, 75, 76, 77, 91, 92, 96, 123], "id": [35, 45, 52, 76, 80, 81, 85, 93, 123], "cudasetdevic": 35, "dump": [36, 52, 96], "base": [36, 50, 58, 63, 64, 66, 71, 72, 76, 82, 88, 90, 91, 93, 97, 101, 102, 107, 111, 116, 122], "stdout": [36, 75], "equivil": 37, "document": [42, 43, 44, 45, 50, 59, 80, 82, 83, 87, 88, 89, 91, 92, 114, 115, 117, 118, 120], "copyright": [42, 43, 44, 45, 83, 89], "c": [42, 43, 44, 45, 52, 59, 64, 67, 70, 71, 72, 75, 76, 83, 90, 96, 100, 114, 115, 117, 120, 123], "corpor": [42, 43, 44, 45], "right": [42, 43, 44, 45, 55, 59, 60, 82, 114, 115, 117], "reserv": [42, 43, 44, 45, 109, 110], "licens": [42, 43, 44, 45, 89], "under": [42, 43, 44, 45, 59, 65, 82, 94, 105, 122], "bsd": [42, 43, 44, 45], "style": [42, 43, 44, 45, 64, 68, 80, 82, 83], "found": [42, 43, 44, 45, 63, 66, 75, 82, 89, 91, 94, 96, 98, 120], "root": [42, 43, 44, 45, 66, 80, 91, 112], "sourc": [42, 43, 44, 45, 54, 59, 64, 65, 67, 71, 72, 73, 74, 75, 76, 77, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114], "tree": [42, 43, 44, 45, 80, 91, 112, 120], "pragma": [42, 43, 44, 45, 91], "onc": [42, 43, 44, 45, 53, 55, 56, 58, 64, 65, 66, 67, 76, 91, 96, 110, 113, 114, 115, 117, 120], "namespac": [42, 43, 44, 45, 51, 55, 69, 76, 91, 93, 96], "ar": [42, 46, 49, 52, 53, 54, 55, 56, 58, 59, 60, 62, 63, 64, 65, 66, 71, 74, 75, 76, 77, 80, 82, 83, 84, 88, 89, 91, 92, 93, 94, 96, 97, 98, 100, 101, 102, 105, 108, 109, 110, 111, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122], "ones": [42, 56, 57, 59, 66, 82, 89, 94, 96, 122], "necessari": [42, 62, 64, 66, 75, 93, 94, 102, 120], "user": [42, 48, 54, 56, 57, 58, 59, 62, 63, 64, 66, 67, 71, 82, 83, 89, 90, 91, 94, 98, 102, 113, 114, 115, 117, 118, 119, 120, 122], "dont": 42, "know": [42, 60, 80, 82, 93, 94, 96, 104], "we": [42, 44, 53, 54, 55, 56, 57, 58, 59, 60, 62, 63, 64, 65, 72, 75, 80, 82, 88, 89, 91, 93, 94, 96, 98, 99, 100, 101, 102, 103, 104, 105, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 121, 122], "want": [42, 56, 65, 66, 67, 68, 72, 88, 89, 91, 92, 94, 96, 102, 103, 114, 115, 117], "use_cmake_generated_export_head": 43, "torch_tensorrt_export": 43, "els": [43, 44, 48, 64, 71, 77, 82, 83, 95, 97, 98, 111, 112], "__gnuc__": 43, "__attribute__": 43, "__visibility__": 43, "hidden": [43, 80], "endif": [43, 44, 45], "doe": [43, 44, 55, 56, 60, 62, 65, 66, 76, 82, 91, 93, 96, 105, 107], "gaurd": 43, "someth": [43, 55, 82, 114, 115, 117], "6": [43, 55, 56, 58, 66, 70, 86, 88, 89, 95, 96, 111], "setup": [43, 67, 91, 114, 115, 117], "alias": 43, "eas": 43, "ts": [43, 52, 56, 68, 69, 76, 88, 89, 90, 92, 118, 121], "torchtrt": [43, 56, 95, 96, 112, 114, 115, 117], "ifndef": [44, 45], "doxygen_should_skip_thi": [44, 45], "get_batch_impl": 44, "element_typ": 44, "super": [44, 88, 93, 94, 96, 103, 111, 112, 118, 119], "batchtyp": 44, "dataloader_": 44, "cache_file_path_": 44, "use_cache_": 44, "auto": [44, 56, 60, 64, 68, 71, 82, 83, 89, 91, 104, 109, 110, 113, 123], "batched_data_": 44, "push_back": [44, 56], "it_": 44, "begin": [44, 65, 66, 82, 103, 108], "hack": 44, "explict": 44, "work": [44, 55, 59, 60, 64, 65, 68, 71, 74, 75, 76, 82, 83, 91, 93, 96, 102, 103, 108, 113, 114, 115, 117, 118], "here": [44, 53, 54, 56, 58, 63, 64, 65, 66, 68, 80, 82, 83, 88, 89, 91, 93, 94, 96, 99, 108, 109, 110, 111, 112, 114, 115, 117, 118, 120, 121], "explic": 44, "just": [44, 45, 55, 56, 64, 65, 69, 73, 75, 82, 84, 88, 89, 90, 92, 93, 96, 98, 100, 116, 120], "still": [44, 56, 65, 66, 91, 94, 103, 122], "static_cast": 44, "option": [44, 48, 52, 56, 57, 59, 62, 63, 64, 65, 71, 75, 76, 77, 82, 86, 91, 94, 96, 97, 98, 103, 104, 106, 108, 119, 120, 121, 123], "batch_siz": [44, 91, 112], "end": [44, 52, 60, 62, 70, 71, 76, 77, 82, 89, 91, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113], "statu": [44, 83], "reset": [44, 97, 98, 103, 107, 120], "incas": 44, "go": [44, 55, 56, 65, 68, 88, 89, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 122], "again": [44, 58, 60, 82, 96, 100], "stringstream": 44, "ss": 44, "cache_": 44, "clear": 44, "ifstream": 44, "io": [44, 67, 114, 115, 117], "binari": [44, 91], "noskipw": 44, "good": [44, 60, 65, 82, 98], "copi": [44, 60, 65, 67, 70, 74, 83, 113], "istream_iter": 44, "back_insert": 44, "nullptr": [44, 45, 49], "ofstream": [44, 89], "cache_fil": [44, 74, 91], "reinterpret_cast": 44, "cache_size_": 44, "arrayref": [45, 48, 49], "friend": 45, "ostream": 45, "os": [45, 67, 98], "dtype": [45, 48, 49, 52, 63, 64, 65, 70, 71, 72, 75, 76, 77, 90, 93, 96, 97, 101, 105, 107, 108, 111, 113, 114, 115, 116, 117, 118, 119], "device_typ": [45, 46, 76, 91, 92, 123], "int64_t": [45, 46, 48, 49, 91, 123], "core": [45, 52, 55, 56, 59, 64, 71, 76, 89, 94, 122, 123], "agx": 45, "platform": [45, 52, 59, 64, 66, 67, 71, 95, 123], "xavier": [45, 123], "dla_cor": [45, 46, 52, 76, 91, 92, 123], "allow_gpu_fallback": [45, 46, 71, 76, 77, 91, 92, 123], "customclasshold": [45, 48], "min_shap": [45, 48, 63, 65, 71, 76, 77, 90, 105, 108, 116, 118], "opt_shap": [45, 48, 63, 71, 76, 77, 90, 105, 108, 116, 118], "max_shap": [45, 48, 63, 65, 71, 76, 77, 90, 105, 108, 116, 118], "shape": [45, 47, 48, 49, 52, 56, 60, 63, 65, 69, 70, 71, 72, 75, 76, 77, 78, 90, 93, 94, 96, 99, 101, 108, 111, 112, 113, 114, 115, 117, 120, 123], "doubl": [45, 48, 49, 52, 63, 71, 76, 77, 82, 120], "tensor_domain": [45, 48, 76], "input_is_dynam": 45, "ivalu": [45, 47, 49, 53, 58, 60, 89], "input_signatur": [45, 47, 49, 77, 90], "nest": [45, 49, 50, 82, 83], "full": [45, 49, 52, 60, 64, 71, 73, 76, 89, 91, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 117, 120, 123], "spec": [45, 48, 49, 52, 73, 76, 77, 92, 98], "flatten": [45, 47, 70, 88, 89, 112], "fixed_s": [45, 49], "reflect": [45, 76], "builderconfig": 45, "graph_input": [45, 49], "enabled_precis": [45, 49, 63, 64, 71, 75, 76, 77, 89, 90, 91, 92, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 109, 110, 111, 112, 113, 114, 115, 117, 119, 123], "disable_tf32": [45, 49, 64, 71, 75, 76, 77, 91, 96, 104, 109, 110], "sparse_weight": [45, 49, 64, 65, 71, 75, 76, 77, 96], "refit": [45, 49, 64, 69, 71, 76, 77, 92, 96, 98, 99, 100, 114], "truncate_long_and_doubl": [45, 49, 63, 64, 77, 106], "allow_shape_tensor": [45, 49, 77], "uint64_t": [45, 49], "num_avg_timing_it": [45, 49, 64, 71, 75, 76, 77, 92, 96], "workspace_s": [45, 49, 52, 64, 71, 75, 76, 77, 96, 102, 105, 107], "dla_sram_s": [45, 49, 52, 64, 71, 75, 76, 77, 96], "1048576": [45, 49, 64, 71, 75, 76, 77, 96], "dla_local_dram_s": [45, 49, 52, 64, 71, 75, 76, 77, 96], "1073741824": [45, 49, 64, 71, 75, 76, 77, 96], "dla_global_dram_s": [45, 49, 52, 64, 71, 75, 76, 77, 96], "536870912": [45, 49, 64, 71, 75, 76, 77, 96], "require_full_compil": [45, 49, 64, 71, 75, 76, 77, 96], "min_block_s": [45, 49, 56, 63, 64, 71, 75, 76, 77, 93, 94, 95, 96, 97, 98, 102, 103, 104, 105, 107, 108, 111, 112], "3": [45, 49, 52, 55, 56, 58, 63, 64, 65, 67, 68, 70, 71, 74, 76, 77, 82, 83, 86, 88, 89, 91, 92, 93, 95, 96, 97, 98, 100, 101, 102, 105, 108, 109, 110, 111, 112, 113, 116, 118, 121, 123], "torch_executed_op": [45, 49, 56, 63, 64, 71, 75, 76, 77, 96, 102, 103, 105, 107, 108], "torch_executed_modul": [45, 49, 56, 71, 76, 77], "member": [46, 47, 48, 49], "hold": [46, 47, 48, 53, 60, 76, 91], "relat": [46, 82, 103, 107], "let": [46, 52, 55, 60, 65, 71, 76, 77, 80, 82, 114, 115, 116, 117, 122], "layer": [46, 49, 52, 53, 55, 60, 62, 64, 65, 71, 75, 76, 77, 89, 91, 94, 96, 109, 110, 112, 114, 115, 116, 117, 118, 119, 122, 123], "thei": [46, 52, 53, 54, 55, 58, 60, 64, 65, 71, 74, 75, 76, 80, 82, 90, 94, 98], "complex": [47, 49, 64, 66, 88, 90, 93, 100, 110], "either": [47, 48, 52, 60, 62, 71, 76, 77, 80, 82, 88, 89, 90, 93, 94, 95, 96, 98, 121], "one": [47, 54, 55, 60, 64, 65, 67, 71, 75, 76, 82, 88, 89, 90, 93, 94, 96, 103, 107, 109, 110, 114, 115, 117], "rang": [48, 49, 52, 65, 76, 93, 96, 97, 98, 101, 104, 105, 113, 116, 118], "optim": [48, 52, 63, 64, 65, 69, 71, 72, 74, 76, 88, 89, 90, 101, 102, 104, 105, 106, 107, 108, 111, 113, 116, 118, 122], "profil": [48, 72, 75, 119], "singl": [48, 52, 55, 56, 65, 76, 82, 88, 89, 91, 108, 111, 113, 120], "repres": [48, 49, 54, 60, 65, 68, 82, 101, 111], "signifi": [48, 55], "static": [48, 49, 53, 60, 63, 64, 71, 76, 77, 80, 89, 101, 112, 118], "three": [48, 57, 59, 65, 72, 76, 82, 83, 114, 115, 116, 117], "min": [48, 52, 60, 70, 76, 98, 104, 105, 118], "optimin": 48, "max": [48, 52, 60, 70, 76, 80, 98, 104, 105, 112, 118], "allow": [48, 49, 52, 53, 54, 55, 56, 62, 64, 65, 66, 71, 76, 77, 80, 93, 94, 96, 98, 101, 102, 105, 107, 108, 113, 120], "argument": [48, 52, 54, 55, 58, 60, 62, 64, 65, 71, 75, 76, 77, 82, 83, 89, 90, 94, 95, 96, 118], "expect": [48, 54, 55, 60, 76, 89, 90, 93, 116], "tradit": [48, 71, 76, 77, 91], "convect": 48, "produc": [48, 53, 54, 58, 60, 63, 76, 82, 89, 116], "low": [48, 65, 94, 100, 111], "high": [48, 55, 56, 80, 94, 96, 122], "weight": [48, 49, 52, 53, 64, 65, 69, 70, 71, 75, 76, 77, 82, 89, 98, 99, 100, 106, 114, 116], "first": [48, 53, 54, 55, 65, 68, 82, 83, 89, 90, 91, 94, 96, 98, 100, 103, 104, 114, 115, 117, 118, 121, 122], "calcul": [48, 53, 56, 89, 96, 113], "detect": [48, 58, 76], "float32": [48, 49, 52, 63, 64, 65, 71, 76, 77, 96, 100, 104, 106, 109, 110, 113, 118, 119], "dynam": [48, 49, 63, 65, 69, 71, 72, 76, 77, 78, 94, 98, 99, 103, 104, 106, 107, 110, 113, 114, 120], "opt": [48, 66, 75, 76, 108], "minimum": [48, 49, 52, 56, 63, 64, 71, 76, 77, 96, 113], "maximum": [48, 49, 52, 64, 65, 71, 72, 76, 77, 104, 105, 107, 113, 114, 115, 117], "accept": [48, 52, 54, 58, 60, 66, 76, 89, 90, 103, 121], "exampl": [48, 56, 58, 59, 60, 65, 66, 71, 73, 75, 76, 77, 78, 80, 81, 83, 86, 88, 89, 90, 91, 93, 94, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 120, 121], "s": [48, 49, 53, 56, 58, 60, 63, 65, 66, 67, 69, 71, 72, 75, 76, 80, 82, 83, 88, 89, 91, 94, 96, 98, 111, 113, 114, 115, 116, 117, 118, 120, 121], "cannot": [48, 55, 56, 65, 66, 71, 75, 76, 77, 81, 88, 93, 95, 96, 101], "through": [48, 53, 54, 55, 56, 58, 64, 65, 71, 73, 74, 82, 89, 90, 96, 100, 102, 116, 122], "altern": [48, 56, 62, 63, 76, 90, 94, 101, 108, 116, 121], "refer": [48, 54, 57, 59, 65, 81, 86, 89, 91, 96, 112, 114, 115, 117, 118, 121], "given": [48, 49, 52, 54, 55, 65, 71, 72, 74, 76, 77, 88, 89, 90, 92, 93, 94, 111, 118], "kernel": [48, 49, 52, 60, 64, 65, 69, 71, 76, 77, 94, 99, 108, 114, 119, 120], "ani": [48, 52, 53, 54, 60, 62, 64, 65, 70, 71, 74, 75, 76, 77, 80, 82, 89, 90, 91, 94, 96, 105, 118], "event": [48, 64, 97, 98], "place": [48, 55, 62, 65, 82, 83, 84, 91, 93, 96, 112], "variabl": [48, 65, 75, 76], "dimens": [48, 55, 65, 72, 76, 105, 116, 118, 119], "domain": [48, 76, 83, 91], "convien": 49, "fix": [49, 65, 82, 93, 96, 120, 123], "describ": [49, 56, 60, 76, 88, 92, 93, 114, 115, 117], "entri": [49, 60, 98], "okai": 49, "ha": [49, 53, 54, 55, 56, 57, 59, 60, 62, 64, 65, 66, 67, 71, 72, 75, 76, 82, 83, 88, 89, 91, 94, 95, 98, 101, 102, 108, 112, 116, 118, 122], "flaten": 49, "precis": [49, 52, 63, 64, 65, 69, 71, 76, 89, 90, 91, 105, 107, 109, 110, 111, 113, 123], "dure": [49, 52, 54, 56, 60, 63, 64, 71, 74, 76, 91, 94, 108, 109, 110, 113, 114, 115, 116, 117, 118, 120], "prevent": [49, 52, 54, 56, 108], "tf32": [49, 52, 64, 71], "comput": [49, 64, 65, 66, 67, 71, 75, 82, 91, 93, 95, 99, 101, 114, 116], "inner": [49, 83, 116], "product": [49, 67, 76], "round": [49, 71, 76, 77, 96], "10": [49, 66, 67, 71, 72, 76, 77, 86, 88, 89, 91, 93, 101, 111, 112, 113, 114, 115, 116, 117, 118, 119], "bit": [49, 60, 65, 66, 71, 76, 77, 89], "mantissa": [49, 71, 76, 77], "befor": [49, 54, 55, 56, 59, 60, 65, 71, 76, 77, 89, 101, 104, 111, 114, 115, 117, 118], "multipli": [49, 71, 76, 77], "accumul": [49, 64, 71, 76, 77, 109, 110, 111], "sum": [49, 65, 70, 71, 76, 77, 96, 112], "23": [49, 55, 71, 76, 77, 83], "behavior": [49, 56, 65, 71, 76, 77, 94, 109, 110, 118, 120, 121], "sparsiti": [49, 52, 65, 71, 76, 77], "conv": [49, 52, 89, 96], "fc": [49, 52, 55], "truncat": [49, 52, 63, 64, 71, 76, 77], "long": [49, 52, 53, 63, 76, 82, 83, 93], "float": [49, 52, 63, 64, 70, 76, 88, 89, 90, 91, 92, 93, 96, 97, 98, 102, 103, 107, 108, 111, 119], "ishap": 49, "restrict": [49, 64, 71, 76, 77, 118], "cuda": [49, 58, 63, 65, 67, 68, 71, 72, 75, 76, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 100, 101, 102, 104, 105, 106, 109, 110, 111, 112, 113, 114, 115, 117, 118, 119, 120, 121], "safeti": [49, 52, 76], "averag": [49, 52, 64, 71, 76, 77, 96], "time": [49, 52, 53, 54, 55, 56, 57, 58, 59, 60, 64, 65, 66, 68, 69, 71, 72, 75, 76, 77, 80, 82, 89, 91, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113], "workspac": [49, 52, 64, 65, 66, 71, 72, 76, 77, 96, 103, 105, 107], "fast": [49, 52, 64, 68, 71, 76, 77], "softwar": [49, 52, 64, 71, 76, 77, 82], "manag": [49, 52, 53, 55, 57, 59, 60, 64, 66, 67, 71, 73, 75, 76, 77, 89, 101, 108, 120], "ram": [49, 52, 64, 71, 76, 77], "commun": [49, 52, 64, 71, 76, 77, 89], "within": [49, 52, 57, 59, 64, 69, 71, 75, 76, 77, 80, 82, 93, 99, 108, 109, 110, 114, 116], "host": [49, 52, 64, 66, 71, 76, 77, 93, 96, 113, 114, 115, 117], "share": [49, 52, 64, 66, 71, 75, 76, 77, 98], "across": [49, 52, 55, 56, 64, 71, 76, 77, 80, 101], "metadata": [49, 52, 54, 58, 60, 64, 71, 76, 77, 80, 102, 118, 119], "quantizatiom": 49, "instead": [49, 52, 53, 54, 55, 66, 71, 75, 76, 89, 94, 102, 111, 112, 120], "potenti": [49, 71, 76, 85, 101], "subgraph": [49, 52, 53, 54, 55, 60, 62, 89, 96, 98, 101, 122], "aten": [49, 54, 55, 56, 60, 61, 64, 69, 70, 71, 76, 77, 89, 94, 103, 108, 122], "thrown": [49, 71, 76, 77], "empti": [49, 71, 72, 76, 77, 83, 88, 96, 111], "torch_tensorrtnamespac": 50, "loggingenum": 50, "levelnamespac": 50, "ptqtemplat": 50, "int8cachecalibratortempl": 50, "int8calibratornamespac": 50, "torchscriptstruct": 50, "compilespecstruct": 50, "deviceclass": 50, "devicetypestruct": 50, "graphinputsstruct": 50, "inputclass": 50, "datatypeclass": 50, "tensorformatenum": 50, "cppdirectori": 50, "includedirectori": 50, "torch_tensorrtfil": 50, "hfile": 50, "relationship": 50, "inherit": [50, 65, 71, 91], "subdirectori": 51, "definit": [51, 54, 60, 82], "cli": [52, 90], "It": [52, 54, 55, 56, 57, 59, 60, 65, 66, 69, 76, 80, 82, 93, 95, 96, 113, 116, 120, 122], "serv": [52, 58, 65, 69, 71, 76], "easi": [52, 53, 55, 89, 91], "wai": [52, 64, 65, 66, 88, 89, 91, 93, 94, 96, 98, 102, 116, 120, 121], "command": [52, 64, 66, 82, 83, 88, 89, 114, 115, 117], "line": [52, 66, 83, 89, 100], "quickli": [52, 89, 91, 114, 115, 117], "part": [52, 56, 59, 65, 75, 80, 81, 82, 93, 96, 98, 101], "deploy": [52, 75, 89, 90, 91, 93, 114, 115, 116, 117, 120, 123], "pipelin": [52, 89, 100, 106, 123], "basic": [52, 56, 65, 83, 114, 115, 117], "featur": [52, 56, 65, 66, 89, 91, 92, 106, 111, 112, 113, 116, 122], "though": [52, 59, 60, 88, 89, 122], "alreadi": [52, 53, 54, 55, 89, 91, 93, 94, 96, 99, 111, 114, 115, 117, 118], "two": [52, 55, 60, 62, 64, 65, 66, 76, 82, 83, 87, 88, 90, 91, 94, 98, 111, 114, 115, 117, 118], "embed": [52, 54, 58, 70, 77, 82, 123], "plan": [52, 59, 63, 64, 71], "after": [52, 53, 55, 56, 62, 65, 71, 75, 76, 88, 89, 90, 101, 103, 107, 114, 115, 117, 120], "link": [52, 53, 62, 69, 80, 81, 86, 89, 96, 120], "against": [52, 89, 94], "libtorchtrt": [52, 66, 89], "python": [52, 56, 59, 62, 64, 65, 67, 71, 72, 75, 76, 77, 82, 83, 89, 92, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 120, 123], "import": [52, 55, 56, 63, 64, 65, 66, 67, 68, 75, 80, 82, 88, 89, 90, 92, 93, 94, 96, 97, 98, 100, 114, 115, 117, 118, 120, 121], "packag": [52, 55, 64, 67, 89], "aspect": 52, "ident": [52, 62, 64, 71, 76, 93, 102], "standard": [52, 58, 66, 69, 71, 75, 76, 77, 82, 92, 93, 94, 96, 100, 111, 116, 120], "load": [52, 56, 58, 64, 65, 68, 71, 74, 75, 76, 77, 89, 90, 91, 92, 93, 95, 96, 97, 98, 100, 102, 104, 113, 114, 115, 116, 117, 120, 122], "like": [52, 53, 55, 58, 60, 65, 66, 68, 76, 81, 82, 88, 89, 90, 91, 93, 94, 96, 98, 100, 102, 104, 113, 114, 115, 117, 120], "would": [52, 54, 60, 64, 65, 66, 67, 75, 89, 90, 92, 94, 96, 104, 114, 115, 117, 120], "input_file_path": [52, 123], "output_file_path": [52, 123], "input_spec": [52, 65, 72], "displai": [52, 62, 64, 73, 80, 120], "menu": [52, 80, 82], "verbios": 52, "v": [52, 67, 83, 112, 114, 115, 117], "verbos": [52, 64, 65, 71, 72, 83, 105, 107], "about": [52, 53, 58, 60, 66, 75, 80, 89, 111, 114, 115, 117, 118], "process": [52, 56, 64, 76, 81, 82, 88, 91, 92, 93, 101, 102, 103, 108, 114, 115, 116, 117, 120], "onto": [52, 58], "consol": 52, "w": [52, 66, 76, 111], "disabl": [52, 64, 66, 71, 75, 80, 81, 94, 98, 113, 120], "i": [52, 55, 60, 66, 68, 70, 82, 83, 88, 89, 91, 93, 96, 97, 98, 101, 104, 109, 111, 112], "debugg": [52, 71, 76, 77], "fallback": [52, 57, 59, 60, 101, 102, 123], "model": [52, 56, 58, 63, 68, 71, 72, 73, 74, 76, 78, 88, 89, 90, 91, 92, 97, 98, 100, 118, 120, 122], "throw": [52, 55, 76, 89], "spars": [52, 54, 64, 70, 71], "p": [52, 70, 89, 114, 115, 117, 123], "repeat": [52, 70], "f32": [52, 71, 75, 76, 96], "half": [52, 64, 76, 82, 89, 90, 91, 92, 96, 101, 103, 105, 109, 110, 111, 113, 119, 123], "float16": [52, 76, 96, 100, 106, 111, 119], "f16": [52, 76, 89, 114, 115, 117, 123], "i8": [52, 76], "d": [52, 67, 76, 82, 83, 89, 123], "multi": [52, 75], "dlacor": 52, "avail": [52, 54, 60, 62, 64, 65, 66, 67, 71, 75, 76, 80, 96, 104, 111, 113, 116, 122, 123], "dla_standalon": [52, 76], "file_path": [52, 76, 95, 121], "teo": 52, "op_nam": 52, "op": [52, 53, 54, 55, 56, 57, 59, 60, 62, 63, 64, 75, 76, 89, 93, 94, 103, 108, 120, 122], "partial": [52, 82], "tem": 52, "module_nam": 52, "mod": [52, 56, 65, 71, 86, 89, 91, 119], "mb": [52, 78], "num_op": 52, "block": [52, 53, 55, 56, 64, 71, 86, 93, 122], "treat": 52, "num": 52, "avg": 52, "num_it": 52, "sram": 52, "local": [52, 55, 66, 67, 80, 89], "dram": 52, "atol": 52, "absolut": [52, 66], "toler": 52, "threshold": 52, "numer": [52, 65, 83], "deviat": 52, "1e": [52, 100, 102], "rtol": 52, "rel": [52, 56, 101], "5": [52, 56, 58, 59, 64, 65, 66, 67, 71, 75, 76, 82, 83, 86, 88, 89, 94, 96, 100, 101, 103, 108, 111, 113, 114, 115, 117], "skip": 52, "complianc": 52, "64bit": [52, 95], "32bit": 52, "custom": [52, 62, 63, 65, 66, 69, 99, 109, 110, 111, 114], "dll": 52, "n": [52, 60, 62, 76, 89, 91, 93, 94, 96, 97], "min_n": 52, "min_c": 52, "min_h": 52, "min_w": 52, "opt_n": 52, "opt_c": 52, "opt_h": 52, "opt_w": 52, "max_n": 52, "max_c": 52, "max_h": 52, "max_w": 52, "32": [52, 76, 88, 89, 90, 91, 104, 109, 110, 112, 123], "flag": [52, 56, 57, 59, 64, 66, 71, 74, 76, 90, 108, 109, 110, 120, 121], "forc": [52, 63, 65, 71, 76, 77, 80, 111], "posit": [52, 54, 65, 76, 80], "test": [52, 56, 59, 65, 66, 67, 71, 76, 82, 83, 91, 112, 114, 115, 116, 117], "ssd_trace": 52, "pt": [52, 65, 89, 104, 109, 110, 114, 115, 117], "ssd_trt": 52, "300": [52, 92, 93], "512": [52, 71, 76, 77, 112, 116], "1024": [52, 71, 76, 77, 93, 109, 116], "simplifi": [53, 96], "form": [53, 75, 76, 82, 90, 114, 115, 117], "up": [53, 55, 56, 57, 58, 59, 62, 65, 66, 71, 76, 82, 88, 93, 94, 96, 98, 101, 102, 103, 107, 108, 113, 116], "context": [53, 57, 58, 59, 64, 73, 75, 94, 101, 104, 108, 111, 120], "inetworkdefinit": [53, 54], "record": [53, 88, 97, 98, 108, 120], "togeth": [53, 60, 89], "start": [53, 56, 65, 70, 74, 76, 83, 89, 92, 96, 97, 98, 104, 116], "look": [53, 54, 55, 68, 71, 76, 88, 91, 92, 94, 98, 104, 114, 115, 117, 118], "assembl": [53, 62, 89], "resourc": [53, 91, 93, 96, 101], "coupl": [53, 59, 65, 120], "state": [53, 54, 60, 62, 75, 89, 94, 100, 104, 111], "been": [53, 60, 64, 66, 67, 74, 83, 89, 95, 98, 101, 102, 111, 122], "evaluated_value_map": [53, 60], "stage": [53, 65], "arg": [53, 54, 62, 65, 71, 74, 75, 76, 86, 89, 94, 95, 96, 98, 112, 116], "itensor": [53, 54, 60, 65, 89, 94, 96], "value_tensor_map": [53, 60], "typic": [53, 60, 76, 101, 108, 114, 115, 117], "abl": [53, 55, 60, 62, 65, 91, 92, 96, 102, 104], "system": [53, 60, 62, 64, 69, 71, 75, 76, 77, 94, 95, 96, 98, 102, 122], "registri": [53, 54, 89, 96], "enter": [53, 76], "recurs": 53, "resolv": [53, 55, 57, 59, 103, 104, 107], "until": [53, 56, 59, 60, 66, 71, 76, 122], "final": [53, 56, 57, 59, 66, 94, 96, 103, 107, 116], "some": [53, 54, 55, 56, 57, 58, 59, 60, 62, 63, 64, 65, 66, 71, 76, 81, 82, 89, 91, 93, 94, 96, 98, 101, 108, 118, 122], "These": [53, 54, 56, 58, 62, 64, 66, 71, 74, 75, 76, 80, 82, 91, 93, 94, 114, 115, 117, 122], "those": [53, 54, 62, 64, 82], "do": [53, 54, 55, 56, 60, 63, 65, 81, 83, 88, 89, 90, 91, 93, 94, 96, 109, 110, 123], "theori": [53, 82], "kind": [53, 65], "common": [53, 55, 65, 72, 82, 94, 98], "prim": [53, 55, 56, 58, 70, 88, 89], "constant": [53, 54, 55, 56, 89, 96, 101], "emit": 53, "listconstruct": [53, 56, 58, 89], "make": [53, 54, 65, 66, 67, 71, 76, 82, 84, 89, 90, 91, 96, 98, 114, 115, 116, 117, 123], "associ": [53, 60, 89, 98, 120], "where": [53, 54, 55, 60, 62, 64, 65, 71, 75, 76, 77, 83, 89, 91, 93, 94, 102, 108], "result": [53, 55, 56, 66, 68, 71, 73, 75, 76, 77, 80, 88, 90, 93, 95, 96, 100, 101, 102, 108, 113, 114, 115, 117, 119, 122], "done": [53, 56, 59, 96, 102, 114, 115, 117, 121], "mai": [53, 54, 56, 58, 59, 65, 66, 71, 75, 76, 77, 82, 83, 88, 89, 90, 91, 94, 96, 102, 103, 107, 108, 113, 114, 115, 117, 120], "For": [53, 56, 62, 63, 64, 65, 66, 68, 72, 76, 80, 82, 83, 88, 89, 91, 92, 93, 94, 96, 100, 103, 112, 114, 115, 116, 117, 120, 121], "more": [53, 64, 65, 66, 67, 69, 71, 76, 80, 83, 88, 89, 90, 91, 92, 96, 98, 100, 101, 105, 107, 111, 114, 115, 117, 120], "writing_convert": [53, 89], "locat": [54, 62, 66, 91, 94, 96], "py": [54, 55, 59, 62, 65, 66, 67, 78, 80, 82, 87, 88, 89, 91, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 117, 118], "convers": [54, 55, 56, 58, 63, 64, 65, 71, 76, 77, 89, 93, 94, 96, 111, 116, 118], "decror": 54, "dynamo_tensorrt_convert": [54, 94, 96], "signatur": [54, 77], "leaky_relu": [54, 70], "def": [54, 62, 65, 82, 88, 90, 93, 94, 96, 97, 98, 101, 103, 108, 111, 112, 113, 114, 115, 117, 118, 119], "leaky_relu_convert": 54, "ctx": [54, 60, 89, 94, 96, 113], "conversionctx": [54, 60, 89, 94], "tupl": [54, 58, 63, 65, 71, 72, 75, 76, 77, 90, 93, 94, 96, 98, 102, 118, 119], "kwarg": [54, 65, 71, 74, 75, 76, 94, 96, 116], "dict": [54, 71, 75, 76, 77, 94, 96, 98], "union": [54, 60, 64, 71, 75, 76, 77, 89, 94], "sequenc": [54, 62, 65, 71, 72, 75, 76, 77, 82, 94, 96, 104, 108, 113, 116], "decor": [54, 62, 65, 94], "kei": [54, 82, 88, 98, 114, 115, 117, 118], "node": [54, 55, 56, 57, 59, 60, 62, 64, 65, 71, 72, 89, 94, 96, 112, 116, 118], "capability_valid": [54, 94], "lambda": [54, 60, 82, 89, 93, 94, 114, 115, 117], "fx": [54, 62, 63, 71, 75, 76, 89, 90, 94, 96, 102, 121], "determin": [54, 55, 64, 65, 76, 93, 94, 113, 118, 120], "properli": [54, 66], "handl": [54, 55, 56, 58, 64, 65, 75, 76, 93, 96, 101, 108], "partition": [54, 71, 76, 96], "sure": [54, 66, 67, 89, 90, 104, 114, 115, 117, 123], "prioriti": [54, 94], "develop": [54, 65, 66, 67, 69, 82, 83, 89, 94, 96], "bodi": [54, 82, 83], "nativ": [54, 59, 61, 89, 93, 94, 96, 102], "numpi": [54, 76, 96, 97, 98, 100, 101, 102, 111, 113, 114, 115, 117], "frozen": 54, "attribut": [54, 55, 56, 58, 65, 76, 82, 89], "previou": [54, 80, 103, 111], "correspond": [54, 60, 65, 66, 75, 76, 94, 98, 100, 104, 112, 120], "edg": [54, 82], "well": [54, 63, 66, 69, 73, 75, 82, 89, 91, 93, 94, 98, 108, 111, 121], "being": [54, 65, 66, 71, 89, 94, 96, 102, 108], "truth": 54, "http": [54, 61, 64, 66, 67, 80, 82, 88, 89, 91, 94, 96, 100, 103, 107, 111, 112, 114, 115, 116, 117, 118, 120], "github": [54, 61, 64, 66, 67, 80, 89, 91, 103, 107, 111, 112, 114, 115, 117, 120], "com": [54, 61, 64, 66, 67, 89, 91, 100, 103, 107, 111, 112, 114, 115, 117, 120], "blob": [54, 61, 66, 80, 91, 98, 111], "main": [54, 55, 56, 57, 58, 59, 60, 63, 65, 66, 80, 82, 84, 89, 94, 96, 109, 111, 112], "src": [54, 58, 61, 70], "native_funct": [54, 61], "yaml": [54, 61], "sinc": [54, 55, 64, 65, 67, 75, 82, 88, 89, 91, 93, 94, 97, 98, 102, 111], "mani": [54, 56, 64, 65, 80, 82, 83, 94, 98, 102, 122], "composit": [54, 89], "raw": [54, 80, 94], "impl": [54, 93, 94], "subpackag": 54, "chain": [54, 60], "primarili": [54, 59, 66, 89, 94], "manipul": [54, 62, 76], "net": [54, 60, 82, 83, 89, 96, 114, 115, 117], "addit": [54, 55, 64, 65, 75, 76, 89, 94, 96, 98, 102, 108, 111, 116, 118], "call_modul": 54, "call_funct": [54, 62, 65], "eg": [54, 114, 115, 117, 119], "aten_": 54, "_leaky_relu": 54, "opoverloadpacket": 54, "while": [54, 56, 66, 75, 91, 94, 100, 101, 113, 114, 115, 116, 117, 120, 122], "opoverload": 54, "particular": [54, 64, 98], "collect": [54, 56, 64, 71, 76, 77, 89, 90, 112], "trtinterpret": [54, 65, 72], "along": [54, 76], "match": [54, 55, 93, 94, 102], "special": [54, 56, 104, 111], "account": [54, 114, 115, 117], "illustr": [54, 65, 104, 105, 109, 110, 111, 116], "scale_grad_by_freq": [54, 70], "embedding_param_valid": 54, "establish": 54, "subset": [54, 64, 71, 76, 91, 116], "converter_util": [54, 96], "enforce_tensor_typ": 54, "dictionari": [54, 76, 77, 92, 103], "between": [54, 55, 56, 60, 66, 76, 82, 83, 91, 93, 98, 100, 109, 113], "possibl": [54, 66, 82, 93, 94, 96, 98, 114, 115, 116, 117], "prefer": [54, 64, 66, 89], "keyword": [54, 62, 71, 75, 76, 77, 94, 103, 107], "both": [54, 56, 64, 66, 69, 71, 72, 75, 76, 80, 82, 88, 91, 94, 96, 98, 114, 115, 117], "enforc": [54, 89], "situat": 54, "partit": [54, 55, 63, 64, 71, 76, 94, 122], "greater": [54, 71, 73, 76], "than": [54, 55, 64, 66, 71, 76, 81, 82, 94, 97, 98, 100, 111, 113, 116, 120], "3d": [54, 65], "autocast": 54, "therebi": [54, 58, 93, 96, 116], "limit": [54, 55, 73, 81, 91, 95, 98, 99, 113, 114, 122], "author": [54, 83], "conv_nod": 54, "7": [54, 56, 58, 59, 75, 76, 82, 86, 89, 93, 96, 103, 105, 107, 112, 118], "ignor": [54, 64, 71, 75, 76, 93, 96], "misc": [54, 93, 96], "trttensor": 54, "np": [54, 94, 96, 97, 98, 100, 101, 102, 111, 113, 114, 115, 117], "ndarrai": [54, 96], "aten_ops_convolut": 54, "conversioncontext": [54, 94, 96], "side": [54, 55, 80, 89, 94], "effect": [54, 55, 64, 65, 71, 80, 89, 91, 94, 96, 101, 116], "term": [54, 76, 82, 83, 91, 93, 94, 96, 116], "getitem": 54, "categor": 54, "modif": [54, 62, 76, 111], "op_evalu": 54, "capbility_valid": 54, "opcod": 54, "decompos": 54, "suboper": 54, "separ": [54, 56, 57, 59, 66], "Such": 54, "via": [54, 64, 65, 67, 69, 71, 75, 76, 77, 80, 86, 90, 91, 103, 105, 107, 109, 110, 111, 116, 118, 120, 121, 122], "register_torch_trt_decomposit": 54, "addmm_replac": 54, "replac": [54, 56, 62, 66, 67, 74, 93, 96, 112, 122], "input_": 54, "mat1": 54, "mat2": [54, 70], "beta": [54, 65, 70, 77], "alpha": [54, 65, 70, 83], "mul": [54, 56, 70, 94, 108], "matmul": [54, 55, 64, 70, 71, 89, 109, 110, 111, 118], "modifi": [54, 56, 62, 65, 83, 100, 118], "edit": [54, 66, 80], "torch_enabled_decomposit": 54, "torch_disabled_decomposit": 54, "disjoint": 54, "preced": [54, 82], "over": [54, 57, 59, 65, 82, 112, 113, 114, 115, 117, 122], "much": [54, 60, 80, 82, 91], "significantli": [54, 55, 80, 93, 98], "easier": [54, 57, 59, 60, 65, 71, 75, 76, 89, 91, 96, 100], "tri": 54, "made": [55, 57, 59, 76, 82], "represent": [55, 60, 65, 88, 104, 111, 116, 122], "instanc": [55, 62, 64, 66, 71, 74, 75, 88, 89, 94, 116, 120], "idea": [55, 82, 94], "reduc": [55, 56, 57, 59, 65, 71, 76, 91, 93, 96, 98, 101, 108, 116, 120], "actual": [55, 58, 60, 65, 88, 89, 96], "aim": [55, 122], "closer": 55, "scope": [55, 96, 103, 107], "csrc": [55, 61], "common_subexpression_elimin": 55, "subexpress": 55, "dead_code_elimin": 55, "exception_elimin": 55, "wa": [55, 58, 62, 64, 65, 71, 75, 76, 82, 89, 94, 95, 122], "1013": 55, "ne": [55, 70], "1012": 55, "24": [55, 67, 114, 115, 117], "lib": [55, 66, 67, 89], "python3": [55, 66, 89], "site": [55, 66, 82, 89], "nn": [55, 61, 65, 71, 72, 75, 76, 77, 88, 89, 90, 93, 94, 96, 103, 108, 111, 112, 118, 119, 122], "batchnorm": 55, "248": 55, "11": [55, 66, 82, 86, 89, 114, 115, 117], "block0": 55, "raiseexcept": 55, "249": 55, "12": [55, 56, 67, 82, 86, 88, 89, 105, 114, 115, 117, 118], "block1": 55, "guard_elimin": 55, "whose": [55, 65, 105], "freeze_modul": 55, "propag": 55, "fuse_addmm_branch": 55, "variant": [55, 120], "caught": 55, "ret": 55, "622": 55, "self": [55, 58, 60, 70, 75, 76, 88, 89, 90, 93, 94, 96, 98, 103, 108, 111, 112, 116, 118, 119, 123], "bia": [55, 70, 89, 112], "x9": 55, "3677": 55, "output0": [55, 114, 115, 117, 119], "add_": [55, 70, 89, 94], "fuse_linear": 55, "back": [55, 56, 58, 59, 75, 76, 82, 88, 89, 93, 96, 122], "fuse_flatten_linear": 55, "implicitli": [55, 76], "connect": [55, 71, 76, 77, 82, 100, 114, 115, 117, 123], "higher": [55, 64, 71, 76, 80, 82, 88, 113], "1d": 55, "lower_graph": 55, "access": [55, 60, 65, 80, 89, 92, 122], "rather": [55, 111], "getattr": [55, 58, 88, 89], "trainabl": 55, "remain": [55, 76, 91, 122], "lower_tupl": 55, "lowersimpletupl": 55, "tupleconstruct": [55, 58], "tupleunpack": 55, "leav": [55, 62, 64, 71], "statement": [55, 82, 94], "loweralltupl": 55, "_all_": 55, "rais": [55, 65, 76, 95], "onnx": 55, "module_fallback": 55, "consist": [55, 65, 82, 96, 101, 108, 111, 120, 122], "pair": [55, 60, 66, 82, 91, 116], "delimit": 55, "around": [55, 58, 60, 64, 66, 71, 75, 82, 85, 88, 96, 101], "second": [55, 65, 82, 90, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113], "mark": [55, 56, 80, 98, 104], "notatemoduleforfallback": 55, "marknodesforfallback": 55, "tell": [55, 56, 57, 58, 59, 60, 82, 93, 122], "them": [55, 56, 58, 63, 64, 65, 66, 71, 75, 80, 89, 93, 96, 98, 110, 111, 116, 118, 122], "peephole_optimz": 55, "intent": [55, 82], "catch": [55, 76, 89], "small": [55, 96, 97, 101, 114, 115, 117], "might": [55, 66, 80, 102, 118], "interest": [55, 82], "now": [55, 56, 59, 60, 65, 66, 76, 82, 89, 92, 93, 94, 96, 98, 102, 113, 119, 120], "expand": [55, 70], "simpli": [55, 103, 116], "remove_contigu": 55, "remove_dropout": 55, "infer": [55, 64, 65, 71, 76, 77, 89, 91, 95, 102, 103, 113, 116, 118, 120, 121, 122], "remove_to": 55, "unpack_addmm": 55, "reus": [55, 65, 71, 91, 93, 98, 101], "dedic": [55, 83], "unpack_log_softmax": 55, "softmax": [55, 65, 70, 112], "loop_unrol": 55, "suffici": [55, 66, 76], "short": [55, 64, 71, 82, 83, 102], "tile_to_repeat": 55, "instruct": [56, 57, 59, 65, 66, 89, 111, 114, 115, 117], "criteria": [56, 57, 59, 64], "lack": [56, 57, 59, 65, 93, 96, 113], "explicitli": [56, 57, 59, 66, 77, 90, 91, 92, 109, 110, 119], "On": 56, "segment": [56, 63, 96, 105, 107, 111, 116], "verifi": [56, 71, 94, 96, 102], "Then": [56, 91, 92, 102, 114, 115, 117], "roughli": [56, 114, 115, 117], "analysi": 56, "everi": [56, 72, 75, 76, 89, 120], "complet": [56, 63, 71, 76, 88, 89], "mean": [56, 60, 64, 65, 70, 71, 72, 103, 113, 114, 115, 117, 122], "trace": [56, 65, 71, 75, 77, 88, 89, 114, 115, 117, 118, 121, 122], "tensorlist": [56, 60], "figur": [56, 83, 85, 111], "our": [56, 59, 63, 88, 89, 114, 115, 117], "stitch": [56, 89], "altogeth": [56, 80], "brief": 56, "descript": [56, 83, 95, 112], "partitioninfo": 56, "api": [56, 59, 60, 62, 63, 64, 65, 75, 76, 77, 81, 89, 90, 91, 92, 96, 103, 104, 105, 108, 113, 114, 115, 116, 117, 118, 120, 121], "maintain": [56, 58, 60, 76, 100, 108, 122], "code": [56, 59, 62, 64, 65, 66, 81, 83, 88, 89, 91, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 118], "mymodel": [56, 63, 68, 90, 93, 96, 118, 121], "ts_model": [56, 89], "trt_model": [56, 92, 96, 105, 109, 110, 111, 112, 113, 114, 115, 117, 121], "off": [56, 58, 108, 111], "consecut": [56, 63], "satisfi": [56, 62, 65], "forced_fallback_op": 56, "randn": [56, 63, 68, 71, 76, 77, 89, 92, 94, 98, 105, 108, 118, 119, 121], "224": [56, 63, 68, 71, 72, 76, 77, 89, 95, 98, 100, 102, 105, 108, 114, 115, 116, 117, 118, 121], "trt_ts_modul": [56, 90], "input_s": 56, "inputrang": 56, "cfg": [56, 89], "relu": [56, 70, 88, 89, 103, 108, 112], "trt_mod": [56, 68, 89, 91, 123], "consid": [56, 77, 89, 96, 119], "segmentmodelwithdependencyawar": 56, "test_segment": 56, "20": [56, 67, 86, 102, 105, 107], "x_lgamma": 56, "lgamma": 56, "y_lgamma": 56, "div": [56, 70], "div_lgamma": 56, "27": [56, 89], "cat": [56, 66, 67, 70, 112, 113], "greedi": [56, 104, 109, 110, 113], "strategi": [56, 76], "travers": [56, 57, 59, 64], "gather": 56, "same": [56, 58, 62, 64, 65, 66, 71, 76, 80, 82, 88, 89, 92, 93, 95, 96, 98, 102, 105, 107, 114, 115, 117, 118, 120, 121], "encount": [56, 64, 66, 94, 103, 104, 107], "4": [56, 58, 63, 64, 65, 66, 70, 76, 78, 80, 82, 83, 86, 89, 96, 103, 106, 107, 108, 112, 118], "suboptim": 56, "arithmet": 56, "split": [56, 65, 70], "own": [56, 60, 64, 66, 71, 82, 89, 98, 112, 114, 115, 117], "could": [56, 64, 65, 96, 105, 107, 120], "rewrit": [56, 62], "portion": [56, 82, 96, 106], "without": [56, 60, 68, 71, 80, 82, 89, 91, 96, 97, 98, 102, 120], "reorder": 56, "seri": 56, "cleanli": 56, "approach": [56, 98], "achiev": [56, 116], "hit": 56, "larger": [56, 71, 76, 80, 113, 116], "boundari": [56, 74, 76], "guarante": [56, 75], "trigger": [56, 64, 65, 76, 89, 98, 100, 102, 122], "appear": [56, 82], "adjac": [56, 71, 76, 82], "As": [56, 65, 66, 76, 89, 93, 94, 96, 98, 102, 108, 122], "clean": [56, 62, 82, 103, 107], "step": [56, 65, 67, 70, 76, 91, 96, 102, 111, 116], "consolid": [56, 88], "further": [56, 64, 65, 120, 122], "merg": 56, "identifi": 56, "do_not_merg": 56, "combin": [56, 64, 65], "condit": [56, 82, 122], "loop": [56, 64, 65, 104, 109, 110], "ir": [57, 59, 60, 63, 64, 68, 71, 76, 88, 89, 90, 99, 101, 103, 105, 107, 108, 114, 118], "larg": [57, 59, 80, 82, 89, 91, 101, 102, 104, 111, 113, 116], "opset": [57, 59, 94], "compon": [57, 59, 66, 67, 74, 88, 120, 122], "evalu": [57, 58, 59, 112], "deploi": [57, 59, 69, 71, 89, 91, 99, 114, 115, 117], "instanti": [57, 58, 59, 60, 89, 106], "wrap": [57, 58, 59, 64, 65, 71, 82, 85, 89, 92, 103, 107, 108], "extend": [57, 59, 60, 70, 89, 98, 116], "providi": [57, 59], "stand": [58, 82], "interpret": [58, 65, 82], "execute_engin": [58, 75, 89], "stack": [58, 70, 91, 112, 122], "machin": [58, 66, 91, 95, 114, 115, 117], "pop": 58, "push": 58, "element": [58, 65, 82, 83, 86, 93], "realiz": 58, "abstract": [58, 60, 83, 94], "__torch__": [58, 88, 89], "portabl": [58, 66, 77], "serializ": [58, 64, 88, 122], "instnanti": 58, "whatev": [58, 65, 96], "self_1": [58, 89], "torchvis": [58, 91, 92, 95, 98, 100, 102, 105, 108, 112, 114, 115, 117], "resnet": [58, 69, 78, 95, 99, 100, 114, 115, 116, 117], "___torch_mangle_4847": 58, "resnet_trt": 58, "input_0": [58, 89], "__torch___torchvision_models_resnet____torch_mangle_4847_resnet_trt_engin": 58, "listunpack": [58, 89], "multipl": [58, 66, 71, 75, 76, 82, 83, 91, 93, 101, 113, 114, 115, 117, 120], "repack": 58, "ssd": 58, "ssd300_trt": 58, "__torch___pytorch_detection_ssd_src_model_ssd300_trt_engin": 58, "holder": [58, 84], "torchbind": 58, "pickler": 58, "seril": 58, "zip": [58, 66, 100, 102, 111, 114], "depickl": 58, "encod": [58, 111, 116], "sm": 58, "correct": [58, 66, 80, 99, 100, 102, 112, 114, 115, 117], "bazel": [59, 66, 67], "linux": [59, 64, 67, 71, 89, 95], "x86_64": [59, 66], "aarch64": 59, "gcc": [59, 89], "untest": 59, "try": [59, 76, 82, 83, 89, 92, 96, 98, 111, 114, 115, 117, 122], "older": 59, "repositori": [59, 66, 80, 87, 111, 114, 115, 117], "notebook": [59, 69, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114], "doc": [59, 61, 66, 67, 80, 81, 82, 87, 94, 96, 118], "docsrc": 59, "third_parti": [59, 66], "toolchain": [59, 66, 67], "unstabl": 59, "subject": [59, 62, 122], "matur": 59, "most": [59, 65, 66, 72, 96, 102, 114, 115, 117, 120, 122], "hood": [59, 105, 122], "major": [59, 65, 76], "top": [59, 80, 84], "coordin": [59, 76, 111], "ingest": 59, "flow": [60, 65, 82, 88, 116], "ilay": 60, "analogu": 60, "goal": [60, 64, 98], "registernodeconversionpattern": [60, 89], "helper": [60, 94], "pattern": [60, 76, 89, 113], "schema": [60, 89, 94, 96], "caus": [60, 64, 80, 103, 105, 107, 113, 120], "acthardtanh": 60, "torchtrt_unus": 60, "hardtanh": [60, 70], "scalar": [60, 70], "min_val": [60, 70], "max_val": [60, 70], "unwraptodoubl": 60, "new_lay": 60, "addactiv": 60, "activationtyp": [60, 65], "kclip": 60, "torchtrt_check": 60, "unabl": [60, 89, 96], "setalpha": 60, "setbeta": 60, "setnam": [60, 89], "util": [60, 62, 74, 77, 89, 91, 103, 107, 109, 110, 111, 112, 113, 114, 115, 116, 117, 122], "node_info": [60, 89], "c_str": [60, 89], "out_tensor": [60, 89], "associatevalueandtensor": [60, 89], "getoutput": [60, 89], "log_debug": 60, "getdimens": [60, 89], "accord": [60, 64, 77, 95], "unwrap": 60, "tool": [60, 64, 65, 66, 89, 94, 98, 116], "don": [60, 65, 80, 82, 83, 91, 94, 112, 114, 115, 117, 118], "annot": [60, 89], "your": [60, 63, 64, 66, 67, 68, 75, 80, 82, 83, 87, 88, 89, 90, 92, 98, 111, 118, 120], "Its": [60, 82], "track": [60, 91], "sort": [60, 70, 92, 111], "live": [60, 82], "directli": [60, 62, 63, 66, 69, 74, 76, 91, 94, 96, 103, 111, 121], "associatevalueandivalu": 60, "inspect": [60, 88, 89], "dataflow": [60, 89], "mechan": [60, 64, 65, 96, 101, 102, 116], "safe": [60, 64, 71, 75, 76, 77, 111], "unsur": 60, "deep": [60, 64, 69, 80, 91, 96, 123], "straight": 60, "chanc": 60, "none": [60, 64, 65, 70, 71, 72, 74, 75, 76, 77, 80, 82, 94, 96, 98, 103, 104, 111, 112, 113], "wrapper": [60, 65, 101, 108, 121], "similar": [60, 63, 64, 65, 66, 89, 92, 93, 96, 109, 110, 111], "tocustomclass": 60, "tensorcontain": 60, "istensor": 60, "iscustomclass": 60, "lot": [60, 63], "singular": 60, "becaus": [60, 65, 66, 72, 88, 89, 93, 94, 96, 97, 98, 101, 108, 113, 119], "alloc": [60, 69, 99, 108, 114], "freed": 60, "destructor": 60, "destroi": [60, 83], "realli": 60, "think": [60, 82], "becom": [60, 66, 100], "benefit": [60, 89, 98, 108, 113], "deal": [60, 98], "quit": [60, 66, 89, 116], "effici": [60, 101, 108, 111], "batch_norm": [60, 70], "fusion": [60, 62, 65], "deeplearn": [61, 65, 67], "sdk": [61, 67, 114, 115, 117, 122], "matrix": 61, "html": [61, 66, 67, 82, 88, 91, 94, 96, 118], "c_api": 61, "python_api": 61, "org": [61, 66, 80, 82, 88, 89, 91, 94, 96, 118, 120], "stabl": [61, 67, 69, 77, 78, 80, 99, 114, 118], "master": [61, 66, 91, 120], "overview": [61, 69, 103, 108], "md": 61, "appli": [62, 63, 91, 102, 104, 108, 111], "desir": [62, 71, 83, 91, 98], "coalesc": 62, "insert": [62, 64, 71, 89, 91, 94, 98, 102], "graphmodul": [62, 63, 71, 72, 76, 89, 90, 96, 102, 121, 122], "caller": 62, "invok": [62, 64, 65, 88, 89, 120], "lint": 62, "recompil": [62, 71, 76, 94, 98, 102, 104, 107, 118, 122], "repair": 62, "disallow": 62, "repair_input_as_output": 62, "gm": [62, 71], "sample_input": [62, 65, 103], "scenario": [62, 64, 100, 101, 113], "clone": [62, 66, 70, 96], "modified_graph": 62, "extract": [62, 89, 111, 116], "placehold": [62, 94], "isinst": [62, 65, 96, 112], "issubclass": 62, "direct": [62, 86, 102, 120], "len": [62, 70, 96, 111], "direct_output": 62, "inserting_aft": 62, "cloned_placehold": 62, "replace_input_with": 62, "date": [62, 83, 122], "eliminate_dead_cod": 62, "logger": [62, 73], "f": [62, 64, 65, 67, 76, 82, 88, 94, 95, 96, 101, 111, 112, 113], "__init__": [62, 75, 76, 82, 88, 93, 94, 96, 98, 103, 111, 112, 118, 119], "pass_manag": 62, "passmanag": 62, "backend": [62, 68, 69, 71, 77, 78, 81, 92, 97, 98, 99, 103, 104, 108, 112, 114, 115, 117, 118], "offer": [62, 64], "registr": [62, 65], "conveni": [62, 91, 107, 116, 120, 122], "control": [62, 65, 88, 102, 113, 120], "_aten_lowering_pass": 62, "my_custom_pass": 62, "front": [62, 71], "passlist": 62, "arbitrari": [62, 75], "remov": [62, 63, 71, 80, 97, 98, 101, 111, 112], "dump_lowering_pass": 62, "apply_lowering_pass": 62, "graph_modul": [62, 71], "_remove_lowering_pass": 62, "evolv": 62, "introduc": [63, 65, 108, 116], "exportedprogram": [63, 68, 71, 76, 102, 109, 110, 113, 118, 122], "dynamo": [63, 64, 66, 68, 74, 75, 76, 78, 89, 93, 94, 95, 96, 97, 98, 99, 101, 102, 103, 105, 107, 108, 112, 113, 114, 115, 117, 118, 119], "frontend": [63, 71, 74, 90, 93, 96, 99, 105, 107, 112, 114, 115, 117, 118], "simpl": [63, 64, 65, 82, 83, 88, 93, 114, 115, 116, 117, 118], "usag": [63, 65, 69, 74, 78, 82, 89, 93, 99, 113, 114, 118, 121], "eval": [63, 68, 89, 90, 94, 95, 97, 98, 100, 101, 102, 103, 104, 105, 107, 108, 109, 110, 111, 112, 113, 114, 115, 117, 118, 119, 121], "exp_program": [63, 98, 102, 111, 112, 118], "trt_gm": [63, 68, 98, 102, 118, 119, 121], "interact": [63, 82, 100, 103, 105, 106, 107, 108], "ideal": 63, "discuss": [63, 64, 114, 115, 117], "section": [63, 65, 80, 82, 83, 84, 86, 89, 91, 114, 115, 117, 121], "frequent": [63, 101], "builder": [63, 64, 65, 71], "respect": [63, 64, 66, 71, 76, 109, 110, 119], "releas": [63, 64, 67, 82], "insid": [63, 82, 93, 96], "decomposit": [63, 64, 71, 76, 93, 96], "downstream": [63, 116], "constraint": [63, 113], "guid": [64, 81], "present": [64, 102], "learn": [64, 66, 69, 89, 91, 96, 114, 115, 117, 123], "acceler": [64, 72, 76, 120, 122, 123], "workflow": [64, 65, 68, 69, 71, 72, 76, 89, 92, 98, 99, 100, 105, 106, 107, 109, 110, 114, 116], "wide": [64, 76, 86], "varieti": [64, 114, 115, 117], "primari": [64, 94, 98, 121], "simplic": 64, "optimized_model": [64, 68, 97, 101, 103, 105, 107], "depth": [64, 80, 116], "challeng": [64, 100, 114, 115, 117], "addition": [64, 96], "fit": [64, 82], "compilationset": [64, 71, 75, 94, 96, 103], "_enum": [64, 71], "callabl": [64, 71, 76], "pass_through_build_failur": [64, 71, 75, 76, 96, 108], "max_aux_stream": [64, 71, 75, 76, 96], "version_compat": [64, 71, 75, 76, 96], "optimization_level": [64, 71, 75, 76, 96, 103], "use_python_runtim": [64, 71, 75, 76, 96, 97, 98, 100, 102, 103], "truncate_doubl": [64, 71, 75, 76, 96, 97, 109, 110, 113], "use_fast_partition": [64, 71, 75, 76, 96], "enable_experimental_decomposit": [64, 71, 75, 76, 96], "_devic": [64, 71], "assume_dynamic_shape_support": [64, 71, 75, 76], "engine_cap": [64, 71, 75, 76, 96], "dryrun": [64, 71, 75, 76, 96], "hardware_compat": [64, 71, 75, 76, 96], "timing_cache_path": [64, 71, 75, 76, 98], "tmp": [64, 71, 75, 76, 89, 97], "torch_tensorrt_engine_cach": [64, 71, 75, 76], "timing_cach": [64, 65, 71, 75, 76], "bin": [64, 66, 67, 71, 75, 76], "lazy_engine_init": [64, 71, 75, 76], "cache_built_engin": [64, 71, 75, 97, 98], "reuse_cached_engin": [64, 71, 75, 97, 98, 102], "use_explicit_typ": [64, 71, 75, 109, 110, 113, 119], "use_fp32_acc": [64, 71, 75, 109, 110, 111], "refit_identical_engine_weight": [64, 71, 75], "strip_engine_weight": [64, 71, 75], "immutable_weight": [64, 71, 75, 76, 97, 98, 100, 102], "enable_weight_stream": [64, 71, 75, 113], "enable_cross_compile_for_window": [64, 71, 75], "use_aot_joint_export": [64, 71, 75], "dpython": [64, 71, 76, 77], "per": [64, 71, 96, 120], "regardless": [64, 71, 83, 105, 107], "fail": [64, 71, 76, 89, 100, 102, 112, 123], "auxiliari": [64, 71], "stream": [64, 69, 71, 76, 93, 96, 99, 114], "impli": [64, 71], "longer": [64, 66, 71, 76, 80, 95, 120], "search": [64, 69, 71, 76, 80], "strictli": [64, 71], "runtim": [64, 66, 68, 69, 71, 76, 89, 94, 99, 100, 103, 107, 108, 113, 114, 122], "presenc": [64, 71, 108], "preferenti": [64, 71], "choos": [64, 65, 71, 88], "float64": [64, 71, 76, 77], "toggl": [64, 71, 76], "mode": [64, 65, 71, 75, 76, 90, 91, 94, 108, 111, 112], "detail": [64, 65, 67, 71, 88, 89, 96, 98, 114, 115, 117, 120], "natur": [64, 71, 82], "architectur": [64, 66, 69, 71, 76, 95, 98, 116], "amper": [64, 71, 76], "newer": [64, 66, 71, 76], "storag": [64, 71, 91], "use_strong_typ": [64, 71], "strong": [64, 71, 82], "mix": [64, 69, 71], "happen": [64, 65, 71, 88, 100, 105, 118], "strip": [64, 71], "non": [64, 66, 71, 76, 83, 85, 111, 120], "refitt": [64, 71, 76, 98], "were": [64, 71, 96, 102, 120], "cross": [64, 71, 82, 99, 114], "window": [64, 71, 82], "aot_export_joint_simpl": [64, 71], "aot_autograd": [64, 71], "distribut": [64, 67, 71, 89, 91, 113, 120], "sub": [64, 70, 82, 88, 103], "slate": 64, "futur": [64, 65, 71, 76, 77, 104, 120], "occur": [64, 108, 113], "first_output": 64, "subsequ": [64, 98, 101, 108], "second_output": 64, "session": [64, 68, 82, 98, 108], "point": [64, 66, 76, 80, 81, 82, 89, 93, 111, 112, 114, 115, 117], "cover": [64, 93, 94], "benchmark": [64, 70], "automat": [64, 67, 76, 82, 89, 99, 102, 114, 118, 122], "vari": [64, 72, 113, 118], "inf": 64, "dynamo_convers": 64, "contribut": [64, 101], "demonstr": [64, 82, 83, 84, 91, 93, 94, 96, 98, 100, 112, 114, 115, 116, 117], "break": [64, 65, 71, 75, 76, 82, 93, 96, 101, 110, 111], "successfulli": [64, 95, 100, 102, 111], "_dynamo": [64, 97, 98, 103, 104, 105, 107, 118], "explain": [64, 65, 69], "veri": [64, 65, 83, 84, 91, 92, 104, 109, 110, 114, 115, 117], "explan": [64, 65], "graph_break_count": 64, "furthermor": 64, "durat": [64, 82], "latter": [64, 75], "logic": [64, 65, 94], "guard": 64, "compos": [65, 88, 91, 94, 112, 114, 115, 117], "variou": [65, 123], "etc": [65, 80, 82, 96, 123], "environ": [65, 68, 71, 114, 115, 117], "research": 65, "few": [65, 66, 76, 94], "nightli": 65, "lower_exampl": 65, "welcom": [65, 89], "finish": 65, "converison": 65, "pleas": [65, 67, 76, 82, 89, 99, 111, 112, 114, 115, 117, 118], "max_batch_s": [65, 72, 114, 115, 117], "2048": [65, 72], "max_workspace_s": [65, 72], "33554432": [65, 72], "explicit_batch_dimens": [65, 72], "lower_precis": [65, 72], "lowerprecis": [65, 72], "verbose_log": [65, 72], "timing_cache_prefix": [65, 72], "save_timing_cach": [65, 72], "cuda_graph_batch_s": [65, 72], "dynamic_batch": [65, 72], "turn": [65, 72, 108], "trtmodul": [65, 72], "otherwis": [65, 66, 72, 98, 120], "implicit": [65, 70, 72, 82], "config": [65, 66, 72, 114, 115, 117], "updat": [65, 66, 67, 71, 72, 76, 96, 102], "dim": [65, 70, 72, 96, 98, 112, 113, 114, 115, 117, 118], "fx2trt_exampl": 65, "acc_trac": 65, "come": [65, 66, 81, 93, 96, 100, 114, 115, 117], "my_pytorch_model": 65, "build_model": 65, "prepar": [65, 114, 115, 117], "acc_mod": 65, "earli": [65, 102], "deprec": [65, 70], "continu": [65, 82, 108, 120], "backward": [65, 75, 96, 122], "vision": [65, 99, 114, 115, 117], "activ": [65, 77, 82, 89, 91, 94, 116, 120, 123], "except": [65, 71, 76], "permut": [65, 70, 111], "transpos": [65, 70, 118], "ll": [65, 98, 104], "inputtensorspec": [65, 72, 76], "experiment": [65, 76, 77], "dataclass": [65, 103], "re": [65, 76, 82, 93, 98, 100, 108, 120], "manual": [65, 76, 81, 82, 102, 113], "sampl": [65, 71, 82, 90, 91, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 114, 115, 117], "rand": [65, 89, 95, 98, 100, 102, 103, 114, 115, 117], "from_tensor": [65, 76], "slightli": [65, 66, 96], "promis": 65, "optimize_target_shap": 65, "input_tensor_spec": 65, "shape_rang": [65, 72], "100": [65, 72, 96, 98, 112, 113], "accordingli": [65, 80, 118, 120], "trtinterpreterresult": [65, 72], "namedtupl": 65, "input_nam": [65, 72], "output_nam": [65, 72], "serialized_cach": [65, 72], "bytearrai": [65, 75, 77], "afford": 65, "temporari": [65, 98], "best": [65, 71, 76, 82, 100, 113, 119], "perforamnc": 65, "examin": 65, "suitabl": [65, 94, 101], "force_fp32_output": 65, "strict_type_constraint": 65, "usual": [65, 66, 80, 104], "unless": 65, "certain": [65, 66, 103, 109, 110, 111, 113, 120], "algorithm_selector": 65, "profiling_verbos": 65, "trt_interpreter_result": 65, "64": [65, 76, 90, 93, 110, 111, 112, 118], "25": [65, 72, 89, 111], "runtimeerror": [65, 112], "xxx": 65, "One": [65, 82, 83, 89, 116, 120], "reload_trt_mod": 65, "reload_model_output": 65, "far": [65, 82], "give": [65, 80, 82], "convtert": 65, "scheme": [65, 71, 76], "action": [65, 82], "tensort": [65, 122], "thing": [65, 66, 82], "compar": [65, 71, 76, 90, 101, 102], "vanilla": 65, "mainli": 65, "builtin": 65, "purpos": [65, 114, 115, 116, 117], "acc_op": 65, "leverag": [65, 91], "power": [65, 82, 89, 113, 116], "goe": [65, 82], "whole": [65, 108], "sigmoid": [65, 70], "tensorrt_convert": 65, "acc_ops_sigmoid": 65, "rest": [65, 82, 83], "input_v": [65, 94], "receiv": 65, "region": 65, "add_activ": 65, "get_output": [65, 96], "wherev": 65, "rememb": [65, 66, 114, 115, 117], "mapper": 65, "todo": [65, 67, 80], "logist": 65, "down": [65, 66, 80, 110], "acc_norm": 65, "foo": [65, 82, 83], "register_acc_op": 65, "register_acc_op_map": 65, "this_arg_is_opt": 65, "op_and_target": 65, "arg_replacement_tupl": 65, "rule": [65, 66, 77], "third": [65, 83], "boolean": [65, 76, 94], "matter": [65, 96], "register_custom_acc_mapper_fn": 65, "design": [65, 74, 94, 100, 109, 113, 116, 123], "redund": 65, "throught": 65, "custom_mapp": 65, "_": [65, 82, 93, 96, 101, 111, 112, 113, 119], "foo_kwarg": 65, "inserting_befor": 65, "foo_nod": 65, "meta": [65, 67, 86, 93, 110, 113], "children": 65, "unit": [65, 76, 108], "test_acc_trac": 65, "acc_op_convert": 65, "essenti": 65, "plugin": [65, 96], "yet": [65, 116], "folder": 65, "center": 66, "pypi": 66, "m": [66, 67, 83, 93, 104, 112], "pip": [66, 67, 99, 104, 111, 114, 115, 117], "upload": [66, 114, 115, 117], "x86": [66, 120], "extra": [66, 75, 89, 96, 100], "url": [66, 80, 114, 115, 117], "download": [66, 67, 86, 91, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 117], "whl": [66, 67], "cu118": 66, "cu124": 66, "tarbal": [66, 89, 91], "easiest": [66, 93, 96], "bazelisk": [66, 67], "bazelbuild": [66, 67], "export": [66, 67, 69, 71, 76, 98, 99, 102, 105, 109, 110, 111, 112, 113, 114, 115, 117, 119, 121, 122], "bazel_vers": 66, "path_to_torchtrt_root": 66, "bazelvers": 66, "mkdir": 66, "cd": [66, 114, 115, 117], "curl": [66, 82], "fssl": 66, "o": [66, 82, 114, 115, 117], "dist": 66, "unzip": 66, "bash": [66, 114, 115, 117], "sh": 66, "cp": [66, 67, 96], "usr": [66, 67], "driver": 66, "branch": [66, 67], "4e5b0f6e860910eb510fa70a76ee3eb9825e7a4d": 66, "l46": 66, "pull": [66, 98, 114, 115, 117], "latest": [66, 67, 80], "l53c1": 66, "fact": 66, "reproduc": 66, "l71": 66, "http_archiv": 66, "build_fil": 66, "archiv": [66, 67], "sha256": 66, "strip_prefix": 66, "OR": 66, "TO": [66, 89], "gnu": 66, "tar": [66, 67, 82, 91], "gz": [66, 82, 83, 91], "ld_library_path": 66, "comment": [66, 82], "uncom": 66, "l114c1": 66, "l124c3": 66, "uv": 66, "astral": 66, "project": [66, 81, 86], "simpler": [66, 91], "wheel": [66, 67], "dep": 66, "lighter": 66, "executor": 66, "avoid": [66, 93, 94, 96, 102, 111, 118], "implic": 66, "python_onli": 66, "legaci": [66, 74], "mainten": 66, "torchdynamo": [66, 118, 122], "technolog": [66, 122], "exclud": [66, 96], "speed": [66, 98, 102], "no_torchscript": 66, "dbg": 66, "pre_cxx11_abi": 66, "complic": 66, "incompat": 66, "popular": [66, 81, 99, 109, 110, 114, 116], "ngc": [66, 67, 114, 115, 117], "tabl": [66, 86], "bdist_wheel": 66, "preinstal": 66, "forum": 66, "correctli": [66, 96], "declar": 66, "intend": [66, 103, 105, 106, 107, 108], "microsoft": 66, "2022": [66, 69], "open": [66, 111, 114, 115, 116, 117], "app": 66, "x64": 66, "prompt": [66, 100, 104, 106, 109, 110, 111], "admin": 66, "privileg": 66, "launcher": 66, "chocolatei": 66, "navig": [66, 80], "ninja": 66, "setuptool": 66, "r": [66, 67, 82, 99, 104, 111, 114], "txt": [66, 67, 99, 104, 111, 114], "distutils_use_sdk": 66, "cuda_win": 66, "libtorch_win": 66, "tensorrt_win": 66, "similarli": [66, 98, 108, 120], "ci_workspac": 66, "win": 66, "tmpl": [66, 67], "torchtrtc": [66, 69, 123], "websit": 66, "finder": 66, "dcmake_module_path": 66, "doesn": [66, 82, 88, 89], "dtorch_dir": 66, "dtensorrt_root": 66, "choic": [66, 74], "b": [66, 70, 76, 83, 93, 113, 114, 115, 117], "dcmake_build_typ": 66, "72048": 66, "jp_workspac": [66, 67], "new_local_repositori": 66, "sudo": [66, 67], "home": 66, "unlik": [66, 92], "libtorch_pre_cxx11_abi": 66, "shift": [66, 70, 82], "jetpack": 66, "jetpack_x": 66, "jetpack_5": 66, "drop": [66, 80, 112], "nvida": 67, "ofjetpack": 67, "With": [67, 80, 82, 89, 91, 93, 98, 114, 115, 117], "incorpor": [67, 83], "cudnn": 67, "9": [67, 86, 89, 95, 96, 114, 115, 117], "dlfw": 67, "09": 67, "jetson": [67, 116], "framework": 67, "instal": [67, 69, 86, 89, 99, 104, 111, 114, 115, 117, 120], "kit": 67, "flash": 67, "board": 67, "apt": 67, "show": [67, 80, 82, 98, 106, 113, 116], "dev": 67, "everth": 67, "nvcc": 67, "cmd": 67, "toolkit": [67, 74], "libcusparselt": 67, "lib64": 67, "wget": [67, 114, 115, 117], "cusparselt": 67, "redist": 67, "libcusparse_lt": 67, "sbsa": 67, "xz": 67, "xf": 67, "v1": [67, 100, 106], "arm64": 67, "mv": 67, "chmod": 67, "pypa": 67, "en": [67, 80], "bootstrap": 67, "jp": 67, "v61": 67, "0a0": 67, "872d972e41": 67, "nv24": 67, "08": [67, 114, 115, 117], "17622132": 67, "cp310": 67, "linux_aarch64": 67, "test_requir": 67, "jetpack6": 67, "lanl": 67, "cuda_vers": 67, "grep": 67, "cut": [67, 82, 102], "sed": [67, 83, 85], "torch_install_path": 67, "dirnam": 67, "__file__": 67, "site_package_path": 67, "cuda_hom": 67, "envsubst": 67, "cxx11": [67, 120], "abi": [67, 120], "anywher": 68, "ahead": [68, 69, 89, 100, 108], "ep": [68, 70, 95, 102, 119, 121], "output_format": [68, 76, 121], "input_tensor": [68, 96, 112, 113], "fill": 68, "aot": [69, 89, 99, 100, 102, 108, 114, 122], "integr": [69, 100, 103], "seamlessli": [69, 76], "ecosystem": [69, 122], "hybrid": [69, 71, 76, 77, 122], "advanc": [69, 78, 83, 91, 99, 104, 114], "bert": [69, 78, 99, 101, 114], "triton": [69, 93, 96], "cudagraph": [69, 99, 114], "overload": [69, 99, 114], "mutabl": [69, 99, 114], "diffus": [69, 78, 99, 114], "gpt2": [69, 99, 114], "llama2": [69, 99, 114], "sam2": [69, 99, 114], "page": [69, 84, 86, 114, 115, 117], "introductori": 69, "blog": [69, 120], "gtc": 69, "2020": [69, 89], "talk": 69, "fall": [69, 76, 93, 96], "2021": 69, "dai": 69, "confer": 69, "_convolut": [70, 89], "stride": [70, 76, 96, 112], "pad": [70, 76, 96, 112], "dilat": 70, "output_pad": 70, "group": [70, 82, 83], "determinist": 70, "cudnn_en": 70, "allow_tf32": 70, "ab": 70, "aco": 70, "acosh": 70, "adaptive_avg_pool1d": 70, "output_s": 70, "adaptive_avg_pool2d": 70, "adaptive_avg_pool3d": 70, "adaptive_max_pool1d": 70, "adaptive_max_pool2d": 70, "adaptive_max_pool3d": 70, "argmax": [70, 113], "keepdim": 70, "argmin": 70, "asin": 70, "asinh": 70, "atan": 70, "atanh": 70, "avg_pool1d": 70, "kernel_s": [70, 96, 112], "ceil_mod": 70, "count_include_pad": 70, "avg_pool2d": 70, "divisor_overrid": 70, "avg_pool3d": 70, "gamma": 70, "var": 70, "momentum": 70, "bitwise_not": 70, "bmm": 70, "ceil": 70, "clamp": 70, "clamp_max": 70, "clamp_min": 70, "constant_pad_nd": 70, "co": [70, 83, 116], "cosh": 70, "cumsum": 70, "tensor_mod": 70, "rounding_mod": 70, "div_": 70, "elu": 70, "scale": [70, 91, 116], "input_scal": 70, "indic": [70, 71, 80, 82, 93, 94, 102, 105, 118, 119], "padding_idx": 70, "eq": [70, 82], "erf": [70, 94], "exp": 70, "expand_a": 70, "fake_quantize_per_channel_affin": 70, "zero_point": 70, "axi": [70, 76, 111], "quant_min": 70, "quant_max": 70, "fake_quantize_per_tensor_affin": 70, "using_int": [70, 89], "start_dim": [70, 89], "end_dim": [70, 89], "floor": 70, "floor_divid": 70, "ge": 70, "gru_cel": 70, "hx": 70, "w_ih": 70, "w_hh": 70, "b_ih": 70, "b_hh": 70, "gt": 70, "hardtanh_": 70, "instance_norm": 70, "running_mean": 70, "running_var": 70, "use_input_stat": 70, "layer_norm": 70, "normalized_shap": 70, "le": 70, "negative_slop": 70, "01": [70, 83, 89, 111, 112], "leaky_relu_": 70, "lstm_cell": 70, "lt": 70, "masked_fil": 70, "mask": [70, 96, 111], "max_pool1d": 70, "max_pool2d": [70, 88, 89], "max_pool3d": 70, "mul_": [70, 94], "narrow": 70, "neg": [70, 100], "norm": 70, "scalaropt_dim": 70, "pixel_shuffl": 70, "upscale_factor": 70, "pow": 70, "tensor_scalar": 70, "expon": 70, "tensor_tensor": 70, "prelu": 70, "prod": [70, 96], "dim_int": 70, "reciproc": 70, "reflection_pad1d": 70, "reflection_pad2d": 70, "relu_": 70, "repeat_interleav": 70, "self_int": 70, "replication_pad1d": 70, "replication_pad2d": 70, "replication_pad3d": 70, "reshap": [70, 96, 111], "roll": 70, "rsub": 70, "scatter": [70, 111], "sigmoid_": 70, "sin": [70, 82], "sinh": 70, "slice": 70, "split_siz": 70, "split_with_s": 70, "sqrt": 70, "squar": 70, "squeez": [70, 111, 116], "sub_": 70, "dim_intlist": 70, "tan": 70, "tanh": [70, 94], "tanh_": [70, 94], "non_block": [70, 112], "memory_format": [70, 76], "prim_devic": 70, "topk": 70, "k": [70, 91, 112], "largest": 70, "dim0": [70, 98], "dim1": 70, "unbind": 70, "unsqueez": [70, 111, 114, 115, 117], "upsample_bilinear2d": 70, "align_corn": 70, "scales_h": 70, "scales_w": 70, "vec": 70, "scale_factor": 70, "upsample_linear1d": 70, "upsample_nearest1d": 70, "upsample_nearest2d": 70, "upsample_nearest3d": 70, "scales_d": 70, "upsample_trilinear3d": 70, "view": [70, 80, 111], "__and__": 70, "__derive_index": 70, "idx": 70, "__getitem__": 70, "__is__": 70, "t1": 70, "t2": 70, "obj": 70, "__isnot__": 70, "__not__": 70, "__or__": 70, "__range_length": 70, "lo": 70, "hi": [70, 82, 83], "__round_to_zero_floordiv": 70, "__xor__": 70, "append": [70, 94, 97, 98, 101, 112, 113], "el": 70, "arang": [70, 93, 96], "pin_memori": 70, "start_step": 70, "copy_": 70, "float_int": 70, "int_float": 70, "floordiv": 70, "is_floating_point": 70, "numel": [70, 93], "l": [70, 112], "9223372036854775807": 70, "requires_grad": 70, "tupleindex": 70, "tup": 70, "exported_program": [71, 76, 121], "arg_input": [71, 76, 94, 102], "kwarg_input": [71, 76, 102], "engine_cache_dir": [71, 97, 98], "engine_cache_s": [71, 97, 98], "5368709120": 71, "custom_engine_cach": [71, 98], "baseenginecach": [71, 98], "int32": [71, 76, 77, 96, 97, 101, 107, 116], "channel_last": [71, 76, 77, 116], "244": [71, 76, 77], "alia": [71, 76], "better": [71, 76, 88, 111, 116, 122], "understand": [71, 76, 118], "convolut": [71, 76, 77, 91, 96, 123], "_c": [71, 76, 77, 92], "oppos": [71, 76, 77], "lean": [71, 76], "spend": [71, 76], "integ": [71, 76, 85], "faster": [71, 76, 97, 98, 101, 111, 116], "parition": [71, 76], "increas": [71, 76, 98, 113], "amount": [71, 76, 113], "defer": [71, 76, 122], "lead": [71, 76, 82, 101, 113, 120], "oversubscript": [71, 76], "hard": [71, 102], "disk": [71, 76, 98], "space": [71, 82, 83, 91], "byte": [71, 75, 76, 77, 96, 98, 113, 116], "1gb": [71, 97, 98], "exce": 71, "oldest": 71, "gear": [71, 91], "toward": [71, 91, 111], "cross_compile_flag": 71, "cross_compil": 71, "refit_module_weight": [71, 102], "compiled_modul": [71, 102], "new_weight_modul": [71, 102], "verify_output": [71, 102], "use_weight_map_cach": [71, 102], "in_plac": [71, 102], "compmil": 71, "coverag": [71, 96], "min_acc_module_s": 72, "is_aten": 72, "use_experimental_fx_rt": 72, "correctness_atol": 72, "correctness_rtol": 72, "minim": [72, 91, 96, 101], "submodul": [72, 88, 96, 108], "fx2trt": 72, "cpu": [72, 101, 109, 110, 111, 113], "has_batch_dim": 72, "dtyep": 72, "prop": 72, "min_input_shap": 72, "optimized_input_shap": 72, "max_input_shap": 72, "popul": 72, "225": [72, 114, 115, 117], "explicit_precis": 72, "logger_level": 72, "model_trt": [73, 93], "model_torchtrt": 73, "internal_error": 73, "dataloadercalibr": [74, 91], "preprocess": [74, 91, 114, 115, 117], "algo_typ": [74, 91], "calibrationalgo": [74, 91], "cachecalibr": [74, 91], "qualnam": [74, 76], "entropy_calibr": 74, "entropy_calibration_2": [74, 91], "legacy_calibr": 74, "minmax_calibr": 74, "set_multi_device_safe_mod": [75, 120], "_multidevicesafemodecontextmanag": 75, "impact": 75, "suppress": 75, "unsaf": 75, "trt_compiled_modul": 75, "torchtensorrtmodul": [75, 96], "encompass": [75, 77], "simpili": 75, "de": 75, "initi": [75, 76, 82, 102, 103, 105, 107, 108, 109, 110], "scriptmodul": [75, 76, 77, 89, 90, 121, 122], "overridden": [75, 76], "subclass": 75, "although": [75, 82], "recip": [75, 91], "afterward": 75, "former": 75, "care": 75, "hook": 75, "silent": 75, "get_extra_st": 75, "state_dict": [75, 76, 100], "set_extra_st": 75, "picklabl": 75, "pickl": [75, 96, 98], "load_state_dict": [75, 100, 112], "pythontorchtensorrtmodul": 75, "serialized_engin": [75, 77], "_set": [75, 103], "weight_name_map": 75, "trt_modul": [75, 120], "engine_str": 75, "my_modul": 75, "current_devic": 75, "disable_profil": 75, "enable_profil": 75, "iprofil": 75, "spent": 75, "get_layer_info": 75, "validate_input_shap": 75, "request": [76, 89, 114, 115, 117], "decid": 76, "deseri": [76, 77, 89, 96], "retrac": 76, "cudagraphstorchtensorrtmodul": 76, "strict": [76, 111, 120], "valueerror": [76, 95], "mutabletorchtensorrtmodul": [76, 100], "pytorch_model": 76, "regular": 76, "whenev": 76, "refit_gm": 76, "shape_mod": 76, "_shapemod": 76, "interv": 76, "notat": 76, "bound": 76, "torch_tensor": 76, "tracer": 76, "example_tensor": 76, "optimization_profile_field": 76, "classmethod": 76, "disable_memory_format_check": 76, "core_id": 76, "schedul": [76, 114, 115, 117], "use_default": 76, "try_to": 76, "anoth": [76, 82, 83, 88, 90, 102], "typeerror": 76, "unknown": 76, "succe": 76, "float_dtyp": 76, "failur": 76, "bf16": 76, "try_from": [76, 96], "complex128": 76, "16": [76, 86, 88, 89, 90, 105, 108], "brain": 76, "bfloat16": 76, "f64": 76, "f8": 76, "fp8": 76, "float8": 76, "i32": 76, "sign": [76, 114, 115, 117], "i64": 76, "u8": 76, "unsign": 76, "uint8": [76, 111], "trt_dla": 76, "torchtrt_dla": 76, "_from": 76, "torchtrt_dla_ec": 76, "torchtrt_safety_ec": 76, "saefti": 76, "trt_dla_ec": 76, "standalon": [76, 82, 111], "certifi": 76, "tf": 76, "torchtrt_linear": 76, "cdhw32": 76, "thirti": 76, "row": [76, 83], "spatial": 76, "31": [76, 89], "subscript": [76, 82], "chw16": 76, "sixteen": 76, "15": [76, 82, 86], "chw2": 76, "chw32": 76, "chw4": 76, "four": [76, 82, 83], "dhwc": 76, "equivi": 76, "channels_last_3d": 76, "dhwc8": 76, "eight": 76, "dla_hwc4": 76, "imag": [76, 91, 96, 100, 106, 112, 114, 115, 117], "roundup": 76, "elements": 76, "dla_linear": 76, "planar": 76, "hwc": 76, "channels_last": 76, "hwc16": 76, "hwc8": 76, "least": [76, 82, 83], "ishapelay": 77, "check_method_op_support": 77, "seriali": 77, "put_binding_nam": 77, "tensorrtcompilespec": [77, 92], "scriptclass": 77, "0x7fe38a1f3730": 77, "_jit_to_tensorrt": 77, "00": 78, "000": [78, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113], "total": [78, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113], "galleri": [78, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114], "mem": 78, "torch_compile_advanced_usag": [78, 103], "torch_compile_resnet_exampl": [78, 105], "torch_compile_stable_diffus": [78, 106], "torch_compile_transformers_exampl": [78, 107], "v0": [79, 114, 115, 117], "pytorch_sphinx_them": [80, 87], "conf": [80, 87], "html_theme_opt": 80, "canonical_url": 80, "analytics_id": 80, "logo_onli": 80, "display_vers": 80, "prev_next_buttons_loc": 80, "bottom": 80, "style_external_link": 80, "vcs_pageview_mod": 80, "collapse_navig": 80, "sticky_navig": [80, 84], "navigation_depth": 80, "includehidden": 80, "titles_onli": 80, "canon": 80, "rank": 80, "trail": 80, "slash": 80, "googl": 80, "analyt": 80, "isn": [80, 82, 93, 96], "shown": [80, 82, 89, 111, 119], "sidebar": [80, 86], "button": [80, 82], "icon": [80, 82], "extern": [80, 82, 99, 114], "display_github": 80, "display_gitlab": 80, "gitlab": 80, "bitbucket": 80, "bar": [80, 82], "www": [80, 82, 89, 91, 114, 115, 117], "sphinx": [80, 81, 82, 83, 87, 93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114], "toctre": 80, "lose": 80, "scroll": [80, 84], "unlimit": 80, "header": [80, 82, 83, 89, 114, 115, 117], "render": 80, "github_url": 80, "bitbucket_url": 80, "gitlab_url": 80, "left": [80, 82], "upon": [80, 103, 107], "rst": [80, 82], "visitor": 80, "revert": 80, "misbuild": 80, "properti": [80, 96], "stick": 80, "screen": 80, "vertic": [80, 82], "too": [80, 82, 83], "sticki": [80, 86], "nav": [80, 86], "At": [81, 94, 102], "django": 81, "payment": 81, "dotpai": 81, "dotpayprovid": 81, "seller_id": 81, "pin": 81, "lock": 81, "lang": 81, "pl": 81, "polish": 81, "gatewai": 81, "transfer": 81, "purchas": 81, "item": [81, 83, 112], "param": 81, "seller": 81, "consult": 81, "ui": 81, "languag": [81, 82, 83, 88, 93, 96, 99, 104, 109, 114, 115, 117], "data_item_1": 81, "emphasi": 82, "hyperlink": 82, "uri": 82, "web": 82, "anonym": 82, "label": [82, 91, 111, 112, 114, 115, 116, 117], "substitut": 82, "charact": 82, "exceedingli": 82, "ugli": 82, "problem": [82, 110], "problemat": 82, "ext": [82, 83], "autodoc": [82, 83], "demo": [82, 91, 111], "test_py_modul": [82, 86], "my": [82, 104, 109], "role": 82, "pep": 82, "287": 82, "rfc": 82, "2822": 82, "superscript": 82, "gui": 82, "taken": [82, 101], "height": [82, 111], "interfer": 82, "press": 82, "keyboard": 82, "mous": 82, "mmb": 82, "menuselect": 82, "seen": [82, 83], "whitespac": 82, "signific": [82, 96], "strang": 82, "hyphen": 82, "word": [82, 116], "adjust": 82, "width": [82, 111, 116], "browser": 82, "sentenc": [82, 113, 116], "suppli": [82, 102], "258": 82, "equat": 82, "x_": 82, "x_0": 82, "x_1": 82, "x_2": 82, "x_3": 82, "x_4": 82, "nabla": 82, "frac": 82, "theta": 82, "phi": 82, "restructuredtext": [82, 83], "parser": [82, 95, 112], "colon": 82, "indent": 82, "literal_block": 82, "spaces_and_linebreak": 82, "preserv": [82, 88, 91, 111], "markup_process": 82, "Or": 82, "great": [82, 89, 93, 96, 98, 122], "why": [82, 120], "didn": 82, "blank": 82, "align": 82, "permit": 82, "awai": 82, "eric": 82, "orchestra": 82, "leader": 82, "bee": 82, "philosoph": 82, "ipso": 82, "facto": 82, "But": [82, 89, 102, 113], "got": [82, 89], "vi": 82, "entiti": 82, "said": 82, "entir": [82, 108, 111, 122], "ancient": 82, "injuri": 82, "sing": 82, "elk": 82, "bracket": 82, "miss": [82, 89], "brontosaurus": 82, "thin": 82, "thicker": 82, "middl": 82, "That": [82, 89], "mine": 82, "belong": 82, "me": [82, 83], "ann": 82, "begun": 82, "past": 82, "pars": [82, 89], "someurl": 82, "dev0": 82, "d6be4ba": 82, "caption": [82, 85], "pane": 82, "shell_command": 82, "echo": 82, "did": 82, "window_nam": 82, "session_nam": 82, "shorthand": 82, "some_funct": 82, "highlight": 82, "THE": 82, "heaven": 82, "hexagram": 82, "six": 82, "unbroken": 82, "primal": 82, "light": [82, 121], "spirit": 82, "weak": 82, "essenc": 82, "energi": 82, "unrestrict": 82, "conceiv": 82, "motion": 82, "regard": [82, 122], "basi": 82, "thu": 82, "persist": 82, "dual": 82, "sens": [82, 89], "univers": 82, "world": 82, "men": 82, "express": 82, "deiti": 82, "human": 82, "denot": [82, 96], "holi": 82, "man": [82, 83], "sage": 82, "ruler": 82, "who": 82, "awaken": 82, "utf": [82, 83], "sphinx_rtd_them": [82, 83], "docstr": [82, 83, 90], "dl": 82, "dt": 82, "tag": [82, 114, 115, 117], "tt": 82, "descnam": 82, "descclassnam": 82, "wrote": 82, "anyth": [82, 83, 111, 120], "programm": 82, "myclass": 82, "dothismethod": 82, "flush": 82, "meth": 82, "capit": 82, "flox": 82, "unreferenc": 82, "nonexist": 82, "extrem": 82, "stuff": 82, "mayb": 82, "bold": 82, "ital": 82, "heck": 82, "put": [82, 93, 116], "13": [82, 86], "backlink": 82, "knowledg": 82, "mind": 82, "ey": 82, "thought": 82, "medium": 82, "peopl": 82, "subsect": 82, "interpol": 82, "indirect": 82, "phrase": 82, "docutil": [82, 83], "sourceforg": [82, 83], "ref": 82, "clickabl": 82, "legend": 82, "revis": [82, 83, 100, 106], "revisit": 82, "enhanc": [82, 101, 111], "structuredtext": 82, "wooden": 82, "nickel": 82, "mad": 82, "scientist": 82, "bigger": 82, "bread": 82, "box": [82, 111, 118, 122], "wash": 82, "behind": 82, "ear": 82, "room": 82, "closet": 82, "bathroom": 82, "trash": 82, "sink": 82, "mother": 82, "g_": 82, "mu": 82, "nu": 82, "pi": 82, "t_": 82, "rho_": 82, "servic": 82, "thing1": 82, "thing2": 82, "thing3": 82, "prose": 82, "provok": 82, "mental": 82, "exert": 82, "reader": 82, "discret": 82, "strongli": [82, 113], "advis": 82, "subtitl": 82, "outsid": 82, "often": 82, "besid": 82, "border": [82, 111], "background": [82, 88], "ok": [82, 89], "transmit": 82, "disconnect": 82, "nonetheless": 82, "semant": 82, "blue": [82, 96], "white": [82, 111], "arab": 83, "roman": 83, "upper": 83, "iii": 83, "iv": 83, "classifi": [83, 88, 89, 112, 116], "paragraph": [83, 86], "z": [83, 93], "commonli": 83, "vm": 83, "david": 83, "goodger": 83, "address": [83, 96, 100], "123": 83, "street": 83, "canada": 83, "a1b": 83, "2c3": 83, "contact": 83, "myself": 83, "organ": 83, "humankind": 83, "2012": 83, "03": 83, "19": [83, 86], "53": 83, "0000": 83, "tue": 83, "jan": 83, "progress": 83, "7302": 83, "wish": 83, "redistribut": 83, "reattribut": 83, "sell": 83, "bui": 83, "rent": 83, "leas": 83, "improv": [83, 101, 108, 111, 120], "quot": 83, "excerpt": 83, "collat": 83, "fold": 83, "stapl": 83, "mutil": 83, "anyon": 83, "heart": 83, "bibliograph": 83, "markup": [83, 86], "literal": 83, "yahoo": 83, "oh": 83, "liter": 83, "heh": 83, "child": 83, "beat": 83, "text": [83, 85, 104, 109, 110, 116], "hehe": 83, "kept": 83, "sai": [83, 116], "cackl": 83, "night": 83, "lone": 83, "guangzhou": 83, "destini": 83, "hope": 83, "dream": 83, "forth": 83, "fifth": 83, "sixth": 83, "lorem": [83, 85], "ipsum": [83, 85], "dolor": [83, 85], "sit": [83, 85], "amet": [83, 85], "consectetur": [83, 85], "adipisc": [83, 85], "elit": [83, 85], "donec": [83, 85], "porttitor": [83, 85], "odio": [83, 85], "posuer": [83, 85], "vita": [83, 85], "ornar": [83, 85], "libero": [83, 85], "matti": 83, "loborti": [83, 85], "justo": [83, 85], "vestibulum": [83, 85], "nibh": [83, 85], "aliquet": [83, 85], "feugiat": [83, 85], "sagitti": [83, 85], "nequ": [83, 85], "qui": [83, 85], "eleifend": 83, "dui": [83, 85], "rutrum": [83, 85], "lectu": [83, 85], "suscipit": [83, 85], "letter": [83, 116], "column": 83, "cell": 83, "span": 83, "nam": [83, 85], "mauri": [83, 85], "arcu": [83, 85], "stub": 83, "behav": 84, "area": 84, "interdum": 85, "nec": 85, "finibu": 85, "dictum": 85, "velit": 85, "ut": 85, "eu": 85, "efficitur": 85, "aliquam": 85, "erat": 85, "diam": 85, "gravida": 85, "imperdiet": 85, "tellu": 85, "nisl": 85, "praesent": 85, "eget": 85, "elementum": 85, "rhoncu": 85, "tincidunt": 85, "suspendiss": 85, "volutpat": 85, "scelerisqu": 85, "tristiqu": 85, "aenean": 85, "condimentum": 85, "risu": 85, "accumsan": 85, "laoreet": 85, "maximu": 85, "sapien": 85, "ligula": 85, "fringilla": 85, "commodo": 85, "proin": 85, "et": 85, "pharetra": 85, "etiam": 85, "turpi": 85, "ant": 85, "luctu": 85, "vel": 85, "malesuada": 85, "dignissim": 85, "mi": 85, "nunc": 85, "augu": 85, "sem": 85, "cursu": 85, "nulla": 85, "pellentesqu": 85, "habit": 85, "morbi": 85, "senectu": 85, "netu": 85, "fame": 85, "ac": 85, "egesta": 85, "placerat": 85, "tortor": 85, "iaculi": 85, "venenati": 85, "cra": 85, "puru": 85, "ero": 85, "vehicula": 85, "fusc": 85, "auctor": 85, "phasellu": 85, "est": 85, "viverra": 85, "conval": 85, "faucibu": 85, "vulput": 85, "feli": 85, "sodal": 85, "maecena": 85, "congu": 85, "semper": 85, "enim": 85, "blandit": 85, "sollicitudin": 85, "urna": 85, "orci": 85, "lacu": 85, "quisqu": 85, "facilisi": 85, "hendrerit": 85, "curabitur": 85, "variu": 85, "bibendum": 85, "massa": 85, "magna": 85, "tempu": 85, "metu": 85, "nisi": 85, "pretium": 85, "leo": 85, "euismod": 85, "ultric": 85, "dapibu": 85, "lacinia": 85, "vivamu": 85, "molesti": 85, "hac": 85, "habitass": 85, "platea": 85, "dictumst": 85, "git": 86, "content": [86, 91, 114, 115, 117], "changelog": 86, "math": 86, "14": [86, 97, 107, 114, 115, 117], "17": 86, "18": [86, 89, 100, 111], "submenu": 86, "symlink": 87, "subtre": 87, "_theme": 87, "html_theme": 87, "html_theme_path": 87, "optimiz": 88, "tutori": [88, 91, 93, 94, 96, 98, 100, 102, 115, 117], "beginn": 88, "intro_to_torchscript_tutori": 88, "briefli": 88, "lenet": [88, 89], "lenetfeatextractor": 88, "conv1": [88, 89], "conv2d": [88, 96, 112], "conv2": [88, 89], "lenetclassifi": 88, "fc1": [88, 89], "120": [88, 89], "fc2": [88, 89], "84": [88, 89], "fc3": [88, 89], "feat": [88, 89, 111], "obvious": 88, "pathwai": 88, "input_data": [88, 90], "traced_model": 88, "pick": [88, 119], "script_model": [88, 92], "perspect": 88, "___torch_mangle_10": 88, "129": 88, "___torch_mangle_9": 88, "119": 88, "___torch_mangle_5": 88, "137": 88, "callmethod": 88, "138": 88, "38": 88, "39": 88, "torch_script_modul": [88, 89], "in_tensor": 88, "fly": 88, "lenet_script": [88, 89], "haven": 89, "acquir": 89, "dyanmo": 89, "almost": [89, 122], "trt_lenet_script": 89, "apr": 89, "56": 89, "04": 89, "credit": 89, "stop": 89, "argc": 89, "argv": 89, "cerr": 89, "cout": 89, "even": [89, 100, 108], "cppdoc": 89, "pretti": 89, "fashion": [89, 116], "enable_precis": 89, "And": 89, "convertgraphtotrtengin": 89, "engine_converted_from_jit": 89, "close": [89, 94, 111], "saw": 89, "576": 89, "346": 89, "539": 89, "0464": 89, "0383": 89, "0678": 89, "0932": 89, "1045": 89, "0805": 89, "0435": 89, "0818": 89, "0208": 89, "0358": 89, "cudafloattyp": 89, "0530": 89, "1691": 89, "2802": 89, "1502": 89, "1056": 89, "1549": 89, "input0": [89, 90], "1063": 89, "input1": [89, 90], "input2": 89, "28": 89, "29": 89, "33": 89, "35": 89, "36": 89, "37": 89, "compilegraph": [89, 91], "transform": [89, 91, 93, 97, 99, 101, 102, 104, 107, 109, 110, 111, 112, 113, 114, 115, 117, 121], "laid": 89, "translat": [89, 102], "aren": 89, "techniqu": [89, 91, 110, 120], "checkmethodoperatorsupport": 89, "modular": 89, "ship": [89, 120], "exhaust": 89, "109": 89, "addlay": 89, "yourself": 89, "question": [89, 94], "outself": 89, "flatten_convert": 89, "unwraptoint": 89, "in_shap": 89, "tovec": 89, "out_shap": 89, "shuffl": [89, 91, 112], "addshuffl": 89, "setreshapedimens": 89, "todim": 89, "extens": [89, 122], "ctype": 89, "cdll": 89, "contributor": 89, "upstream": 89, "pr": 89, "usecas": 90, "sole": [90, 91, 122], "individu": 90, "accuraci": [91, 111, 116], "loss": [91, 116], "infrastructur": [91, 114, 115, 117], "streamlin": [91, 93], "expos": [91, 96], "cpp_frontend": 91, "loading_data_recip": 91, "cifar10": [91, 112], "cstddef": 91, "ktrain": 91, "ktest": 91, "un": 91, "cs": 91, "toronto": 91, "edu": 91, "kriz": 91, "cifar": 91, "is_train": 91, "trim": 91, "use_subset": 91, "new_siz": 91, "mode_": 91, "images_": 91, "targets_": 91, "calibration_dataset": 91, "data_dir": 91, "320": 91, "4914": [91, 112], "4822": [91, 112], "4465": [91, 112], "2023": [91, 112], "1994": [91, 112], "2010": [91, 112], "dataloaderopt": 91, "worker": 91, "virtual": 91, "input_shap": [91, 123], "compile_spec": [91, 95, 105, 123], "kf16": [91, 123], "ki8": 91, "vgg16": [91, 112], "testing_dataset": [91, 112], "totensor": [91, 112, 114, 115, 117], "testing_dataload": [91, 112], "num_work": [91, 112], "vgg": [91, 112], "test_ptq_dataloader_calibr": 91, "test_ptq_trt_calibr": 91, "krizhevski": 91, "hinton": 91, "2009": 91, "tini": 91, "simonyan": 91, "zisserman": 91, "2014": 91, "recognit": [91, 116], "arxiv": 91, "preprint": 91, "1409": 91, "1556": 91, "_jit_to_backend": 92, "mobilenet_v2": 92, "pretrain": [92, 98, 100, 104, 105, 108, 114, 115, 116, 117], "cost": [93, 96, 98, 102, 120], "perhap": [93, 96], "overhead": [93, 96, 101, 108, 113, 120], "involv": [93, 101, 102, 108], "greatli": 93, "perviou": 93, "elementwis": [93, 94], "launch": [93, 96, 108, 114, 115, 117], "tensorrt_bind": 93, "trtp": 93, "tl": [93, 96], "elementwise_mul_kernel": 93, "block_siz": [93, 96], "thread": [93, 120], "pid": [93, 96], "program_id": [93, 96], "block_start": 93, "offset": 93, "x_val": 93, "y_val": 93, "wise": 93, "z_val": 93, "custom_op": [93, 96], "torchtrt_ex": [93, 96], "elementwise_mul": 93, "mutates_arg": [93, 96], "assert": [93, 96, 100, 102], "is_cuda": 93, "empty_lik": 93, "grid": 93, "cours": 93, "register_fak": [93, 96], "creation": 93, "less": 93, "boilerpl": [93, 94], "tensordesc": 93, "prior": [93, 94, 98, 118, 120], "x_t": 93, "as_tensor": [93, 96, 111], "y_t": 93, "z_t": 93, "generate_plugin_convert": 93, "supports_dynamic_shap": [93, 94], "my_model": [93, 96], "allclos": [93, 94, 100, 102], "ran": 93, "minut": [93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113], "auto_generate_convert": 93, "jupyt": [93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114], "ipynb": [93, 94, 95, 96, 97, 98, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113], "gelu": 94, "sy": 94, "approxim": 94, "suppos": 94, "my_mod": 94, "ex_input": [94, 96], "baselin": 94, "my_standard_gelu": 94, "supersed": 94, "converterprior": 94, "vers": 94, "distinct": 94, "prepend": 94, "candid": 94, "primit": 94, "compiler_ir": 94, "focu": [94, 100], "interoper": 94, "aten_ops_gelu": 94, "sourceir": 94, "cheap": 94, "unqiu": 94, "op_count": 94, "get_op_count": 94, "nonloc": 94, "source_ir": 94, "lhs_val": 94, "rhs_val": 94, "x_7": 94, "x_8": 94, "79788456080000003": 94, "x_9": 94, "044714999999999998": 94, "x_10": 94, "x_11": 94, "x_12": 94, "x_13": 94, "x_14": 94, "x_15": 94, "my_custom_gelu": 94, "my_mod_erf": 94, "my_gelu_erf": 94, "notic": [94, 101], "converter_overload": 94, "geforcertx": 95, "4080": 95, "3080": 95, "cross_runtime_compilation_for_window": 95, "trt_resnet": 95, "argpars": [95, 112], "argumentpars": [95, 112], "comil": 95, "add_argu": [95, 112], "parse_arg": [95, 112], "manual_se": [95, 97, 98, 100, 102], "resnet18": [95, 98, 100, 102, 105, 108], "amd64": 95, "loaded_model": 95, "load_cross_compiled_exported_program": 95, "trt_output": 95, "cross_compile_for_window": 95, "sake": 96, "circular": 96, "red": [96, 111], "green": [96, 111], "twice": 96, "written": 96, "openai": 96, "formal": 96, "circ_pad_kernel": 96, "all_pads_0": 96, "all_pads_2": 96, "all_pads_4": 96, "all_pads_6": 96, "orig_dims_0": 96, "orig_dims_1": 96, "orig_dims_2": 96, "orig_dims_3": 96, "y_shape_1": 96, "y_shape_2": 96, "y_shape_3": 96, "x_len": 96, "y_len": 96, "mask_i": 96, "i3": 96, "i2": 96, "i1": 96, "i0": 96, "j0": 96, "j1": 96, "j2": 96, "j3": 96, "load_idx": 96, "mask_x": 96, "triton_circular_pad": 96, "out_dim": 96, "tolist": 96, "all_pad": 96, "zero": 96, "orig_dim": 96, "blocksiz": 96, "256": [96, 111, 112, 113, 114, 115, 117], "numblock": 96, "tracabl": 96, "prerequisit": 96, "fake": 96, "real": 96, "faketensor": 96, "autograd": 96, "beyond": 96, "register_autograd": 96, "padded_x": 96, "2604": 96, "4232": 96, "3041": 96, "0833": 96, "2461": 96, "1270": 96, "2450": 96, "4079": 96, "2887": 96, "2828": 96, "0373": 96, "0332": 96, "3143": 96, "6344": 96, "5638": 96, "1867": 96, "5068": 96, "4363": 96, "7937": 96, "3488": 96, "1350": 96, "7966": 96, "3517": 96, "1379": 96, "5537": 96, "1088": 96, "8950": 96, "0550": 96, "6163": 96, "0109": 96, "5245": 96, "9632": 96, "5686": 96, "3775": 96, "8162": 96, "4216": 96, "4311": 96, "1649": 96, "2091": 96, "3668": 96, "1006": 96, "1447": 96, "0352": 96, "7689": 96, "8131": 96, "_run_on_gpu_0": 96, "_run_on_acc_1": 96, "dry": 96, "50": [96, 116], "count": 96, "__": 96, "aggreg": 96, "stat": 96, "latenc": [96, 111, 113, 120], "abstractli": 96, "pkl": [96, 100], "cupi": 96, "gap": 96, "prealloc": 96, "circularpaddingplugin": 96, "ipluginv2dynamicext": 96, "field_collect": 96, "pluginfieldcollect": 96, "x_shape": 96, "num_output": 96, "plugin_namespac": 96, "plugin_typ": 96, "plugin_vers": 96, "get_output_datatyp": 96, "input_typ": 96, "get_output_dimens": 96, "output_index": 96, "dimsexpr": 96, "exprbuild": 96, "iexprbuild": 96, "output_dim": 96, "dimensionoper": 96, "configure_plugin": 96, "inp": 96, "dynamicplugintensordesc": 96, "x_dim": 96, "desc": 96, "supports_format_combin": 96, "po": 96, "in_out": 96, "plugintensordesc": 96, "num_input": 96, "enqueu": 96, "input_desc": 96, "output_desc": 96, "in_dtyp": 96, "a_mem": 96, "unownedmemori": 96, "items": 96, "c_mem": 96, "a_ptr": 96, "memorypoint": 96, "c_ptr": 96, "a_d": 96, "memptr": 96, "c_d": 96, "a_t": 96, "c_t": 96, "cloned_plugin": 96, "__dict__": 96, "circularpaddingplugincr": 96, "iplugincr": 96, "field_nam": 96, "pluginfield": 96, "pluginfieldtyp": 96, "create_plugin": 96, "pluginfieldcollection_": 96, "deserialize_plugin": 96, "pads_dict": 96, "creator": 96, "trt_plugin_registri": 96, "get_plugin_registri": 96, "register_cr": 96, "untyp": 96, "get_trt_tensor": 96, "set_layer_nam": 96, "recal": 96, "intlist": 96, "circular_padding_convert": 96, "retriev": 96, "elsewher": 96, "plugin_registri": 96, "plugin_cr": 96, "get_plugin_cr": 96, "field_config": 96, "eventu": 96, "freez": 96, "_input": 96, "add_plugin_v2": 96, "circular_padding_plugin": 96, "_run_on_acc_0": 96, "grad_fn": 96, "subbackward0": 96, "custom_kernel_plugin": 96, "engine_caching_exampl": [97, 98], "remove_timing_cach": [97, 98], "bertmodel": [97, 101, 107], "random": [97, 98, 100, 102, 111, 113], "seed": [97, 98, 100, 102], "from_pretrain": [97, 100, 101, 104, 106, 107, 109, 110, 111, 113], "uncas": [97, 101, 107, 116], "return_dict": 97, "randint": [97, 101, 107, 113], "compile_bert": 97, "enable_tim": [97, 98], "1st": [97, 98], "measur": [97, 98, 113], "2nd": [97, 98], "3rd": [97, 98], "slower": [97, 98], "messur": [97, 98], "compilation_kwarg": [97, 107], "torch_trt_bert_engine_cach": 97, "30": [97, 98, 100, 102, 105, 107, 111, 119], "synchron": [97, 98, 101, 113], "elapsed_tim": [97, 98], "millisecond": 97, "__name__": [97, 103, 107], "__main__": [97, 103, 107], "engine_caching_bert_exampl": 97, "paid": 98, "upfront": 98, "invalid": 98, "repeatedli": 98, "mitig": [98, 101], "explor": 98, "torch_trt": [98, 100, 102], "_default": 98, "_engine_cach": 98, "flexibl": [98, 122], "histor": 98, "barrier": 98, "reconstruct": 98, "ti": 98, "hash": 98, "magnitud": 98, "torch_compil": [98, 103, 105, 107, 108, 118, 122], "compiled_model": 98, "ms": [98, 101, 113], "dynamo_compil": 98, "example_input": 98, "200": 98, "dynamic_shap": [98, 118], "remot": 98, "systen": 98, "agnost": 98, "implent": 98, "ramenginecach": 98, "held": 98, "engine_cach": 98, "torch_compile_my_cach": 98, "_torch_compile_gpt2": [99, 114], "_torch_export_gpt2": [99, 114], "_torch_export_llama2": [99, 114], "_torch_export_sam2": [99, 114], "sphx_glr_tutorials__rendered_examples_dynamo_cross_runtime_compilation_for_window": [99, 114], "straightforward": 100, "especi": [100, 101], "hug": [100, 104, 109, 110], "face": [100, 104, 109, 110], "difficult": 100, "ever": [100, 104], "walk": [100, 102, 104, 109], "lora": [100, 102], "use_python": 100, "mutable_modul": 100, "model2": [100, 102], "expected_output": [100, 102], "refitted_output": [100, 102], "reload": [100, 122], "checkpoint": [100, 112], "civitai": 100, "12597": 100, "moxin": 100, "diffusionpipelin": [100, 106], "no_grad": [100, 101, 104, 109, 110, 112, 113], "model_id": [100, 106], "runwayml": 100, "hous": 100, "forest": 100, "shuimobysim": 100, "wuchangshuo": 100, "qualiti": 100, "worst": 100, "lowr": 100, "cloudi": 100, "watermark": 100, "pipe": [100, 106], "torch_dtyp": [100, 106], "unet": [100, 106], "negative_prompt": 100, "num_inference_step": 100, "without_lora_mut": 100, "jpg": [100, 111, 114, 115, 117], "procedur": 100, "load_lora_weight": 100, "stablediffusionapi": 100, "load_lora_embed": 100, "weight_nam": 100, "safetensor": 100, "adapter_nam": 100, "lora1": 100, "set_adapt": 100, "adapter_weight": 100, "fuse_lora": 100, "unload_lora_weight": 100, "with_lora_mut": 100, "mutable_torchtrt_module_exampl": 100, "act": 101, "concurr": [101, 114, 115, 117], "overlap": 101, "particularli": 101, "cycl": 101, "overal": [101, 116], "workload": 101, "enough": 101, "overshadow": 101, "cumul": 101, "priorit": 101, "comprehens": 101, "infrequ": 101, "timeit": [101, 113], "test_module_perf": 101, "warm": [101, 108, 113], "accur": 101, "start_tim": [101, 113], "default_tim": [101, 113], "end_tim": [101, 113], "time_m": 101, "median": 101, "metric": 101, "128": [101, 111, 112, 113], "enable_pre_allocated_output": 101, "out_trt": [101, 108], "pre_allocated_output_ctx": 101, "set_pre_allocated_output": 101, "time_opt": 101, "time_norm": 101, "time_opt_m": 101, "1000": [101, 112, 113, 114, 115, 117], "time_normal_m": 101, "3f": [101, 111], "pre_allocated_output_exampl": 101, "expens": 102, "occasion": [102, 103, 107], "adapt": 102, "infeas": 102, "focus": 102, "mostli": 102, "recogn": 102, "behalf": 102, "init": [102, 112], "sett": 102, "randomli": 102, "exp_program2": 102, "compiled_trt_ep": 102, "new_trt_gm": 102, "accomplish": 102, "gaurente": 102, "attempt": [102, 112, 118], "rebuild": 102, "heurist": 102, "refit_engine_exampl": 102, "x_out": 103, "y_out": 103, "x_y_out": 103, "invoc": 103, "sample_inputs_half": 103, "model_half": 103, "backend_kwarg": 103, "optimized_model_custom": 103, "exit": [103, 107], "2052": [103, 107], "compile_engine_and_inf": [103, 107], "art": [104, 111], "causal": 104, "unidirect": 104, "corpu": [104, 116], "huggingfac": [104, 109, 110, 116], "automodelforcausallm": [104, 109, 110, 113], "autotoken": [104, 109, 110], "success": 104, "max_length": 104, "token": [104, 109, 110, 116], "kv_cach": [104, 109, 110], "pad_token_id": [104, 109], "eos_token_id": [104, 109, 110], "attn_implement": [104, 109, 110, 113], "eager": [104, 109, 110, 113], "enjoi": [104, 109], "cute": [104, 109], "dog": [104, 109], "model_input": [104, 109, 110], "return_tensor": [104, 109, 110], "input_id": [104, 109, 110], "regress": [104, 109, 110], "pyt_gen_token": [104, 109, 110], "mark_dynam": [104, 105, 118], "1023": 104, "trt_gen_token": [104, 109, 110], "skip_special_token": [104, 109, 110], "torch_compile_gpt2": 104, "new_input": [105, 107], "new_output": [105, 107], "new_batch_size_input": 105, "new_batch_size_output": 105, "inputs_bs8": 105, "outputs_bs8": 105, "No": [105, 118], "inputs_bs12": 105, "outputs_bs12": 105, "compvi": 106, "majest": 106, "castl": 106, "cloud": 106, "majestic_castl": 106, "png": [106, 111], "enable_cudagraph": [108, 120], "cudagraphs_modul": 108, "set_cudagraphs_mod": [108, 120], "inputs_2": 108, "inputs_3": 108, "out_trt_2": 108, "out_trt_3": 108, "diminish": 108, "encapsul": 108, "wrapped_modul": 108, "captur": 108, "replai": 108, "samplemodel": 108, "intention": 108, "Of": 108, "manner": 108, "opt_with_graph_break": 108, "torch_export_cudagraph": 108, "export_llm": [109, 110, 113], "max_token": [109, 110, 113], "gpt2_ep": 109, "max_seq_len": [109, 110, 113], "parallel": 109, "paradigm": 109, "torch_export_gpt2": 109, "llama_path": [110, 113], "llama": [110, 113], "7b": [110, 113], "chat": [110, 113], "hf": [110, 113], "llama2_ep": [110, 113], "batch_decod": 110, "clean_up_tokenization_spac": 110, "solv": [110, 111, 114, 115, 117], "smaller": [110, 116], "subproblem": 110, "torch_export_llama2": 110, "foundat": 111, "promptabl": 111, "video": 111, "fork": 111, "condition": 111, "concaten": 111, "layernorm": 111, "reli": 111, "stabil": 111, "matplotlib": 111, "pyplot": 111, "plt": 111, "panda": 111, "pd": 111, "pil": [111, 114, 115, 117], "sam2_image_predictor": 111, "sam2imagepredictor": 111, "sam_compon": 111, "sam2fullmodel": 111, "agg": 111, "facebook": 111, "hiera": 111, "set_imag": 111, "predict": 111, "predictor": 111, "image_encod": 111, "forward_imag": 111, "_prepare_backbone_featur": 111, "directly_add_no_mem_emb": 111, "no_mem_emb": 111, "_featur": 111, "prompt_encod": 111, "sam_prompt_encod": 111, "mask_decod": 111, "sam_mask_decod": 111, "_bb_feat_siz": 111, "point_coord": 111, "point_label": 111, "backbone_out": 111, "vision_feat": 111, "feat_siz": 111, "image_emb": 111, "high_res_feat": 111, "high_res_featur": 111, "feat_level": 111, "sparse_embed": 111, "dense_embed": 111, "low_res_mask": 111, "iou_predict": 111, "image_embed": 111, "image_p": 111, "get_dense_p": 111, "sparse_prompt_embed": 111, "dense_prompt_embed": 111, "multimask_output": 111, "repeat_imag": 111, "sam_model": 111, "input_imag": 111, "truck": 111, "rgb": 111, "sam2transform": 111, "facebookresearch": 111, "preprocess_input": 111, "orig_hw": 111, "_transform": 111, "500": 111, "375": 111, "unnorm_coord": 111, "transform_coord": 111, "postprocess": 111, "plot": 111, "confid": 111, "score": 111, "postprocess_mask": 111, "resolut": [111, 116], "sorted_indic": 111, "argsort": 111, "show_mask": 111, "ax": 111, "random_color": 111, "255": 111, "144": 111, "astyp": 111, "mask_imag": 111, "cv2": 111, "contour": 111, "findcontour": 111, "retr_extern": 111, "chain_approx_non": 111, "smooth": 111, "approxpolydp": 111, "epsilon": 111, "drawcontour": 111, "thick": 111, "imshow": 111, "show_point": 111, "coord": 111, "marker_s": 111, "pos_point": 111, "neg_point": 111, "marker": 111, "edgecolor": 111, "linewidth": 111, "visualize_mask": 111, "title_prefix": 111, "overlaid": 111, "figsiz": 111, "gca": 111, "titl": 111, "fontsiz": 111, "savefig": 111, "_output_mask_": 111, "snippet": 111, "torchtrt_input": 111, "unnormalized_coordin": 111, "foreground": 111, "trt_out": 111, "trt_mask": 111, "trt_score": 111, "sam": 111, "torch_export_sam2": 111, "modelopt": 112, "mtq": 112, "export_torch_mod": 112, "layer_spec": 112, "num_class": 112, "init_weight": 112, "in_channel": 112, "pool": [112, 123], "maxpool2d": 112, "batchnorm2d": 112, "sequenti": 112, "avgpool": 112, "adaptiveavgpool2d": 112, "4096": 112, "dropout": 112, "_initialize_weight": 112, "kaiming_normal_": 112, "fan_out": 112, "nonlinear": 112, "constant_": 112, "elif": 112, "normal_": 112, "vgg16_cfg": 112, "ckpt": 112, "model_state_dict": 112, "device_count": 112, "ordereddict": 112, "new_state_dict": 112, "forget": 112, "training_dataset": 112, "randomcrop": 112, "randomhorizontalflip": 112, "training_dataload": 112, "drop_last": 112, "crit": 112, "crossentropyloss": 112, "calibrate_loop": 112, "pred": 112, "5f": 112, "acc": 112, "2f": 112, "quantize_typ": 112, "quant_cfg": 112, "int8_default_cfg": 112, "fp8_default_cfg": 112, "forward_loop": 112, "qdq": 112, "incomplet": 112, "functionaltensor": 112, "functionaltensormod": 112, "_trace": 112, "_export": 112, "float8_e4m3fn": 112, "class_prob": 112, "class_pr": 112, "test_prob": 112, "test_pr": 112, "test_loss": 112, "test_acc": 112, "vgg16_ptq": 112, "overcom": 113, "throughput": 113, "sometim": [113, 118], "outweigh": 113, "slowdown": 113, "hardwar": [113, 123], "experi": 113, "balanc": 113, "time_gener": 113, "output_seq_length": 113, "seq_len": [113, 118], "llm": 113, "input_seq": 113, "inputs_copi": 113, "decod": 113, "logit": 113, "next_token_logit": 113, "next_token": 113, "time_mean_m": 113, "isl": 113, "osl": 113, "solut": 113, "insight": 113, "weight_streaming_ctx": 113, "weight_stream": 113, "mean_lat": 113, "percentag": 113, "weight_budget_pct": 113, "device_budget": 113, "total_device_budget": 113, "permiss": 113, "equal": 113, "proportion": 113, "streamabl": 113, "streamable_budget": 113, "requested_budget": 113, "get_automatic_weight_streaming_budget": 113, "weight_streaming_exampl": 113, "hand": [114, 115, 117], "consider": [114, 115, 117], "grpc": [114, 115, 117], "aforement": [114, 115, 117], "familiar": [114, 115, 117], "resnet50": [114, 115, 117], "torchhub": [114, 115, 117], "docker": [114, 115, 117], "login": [114, 115, 117], "xx": [114, 115], "yy": [114, 115, 117], "mm": [114, 115, 117], "publish": [114, 115, 117], "pwd": [114, 115, 117], "scratch_spac": [114, 115, 117], "nvcr": [114, 115, 117], "py3": [114, 115, 117], "hub": [114, 115, 117], "_validate_not_a_forked_repo": [114, 115, 117], "ts_trt_model": [114, 115, 117], "triton_exampl": [114, 115, 117], "model_repositori": [114, 115, 117], "rm": [114, 115, 117], "highli": [114, 115, 116, 117], "suggest": [114, 115, 117], "simplest": [114, 115, 117], "pbtxt": [114, 115, 117], "data_typ": [114, 115, 117], "type_fp32": [114, 115, 117], "exact": [114, 115, 117], "encourag": [114, 115, 117], "proce": [114, 115, 117], "8000": [114, 115, 117], "8001": [114, 115, 117], "8002": [114, 115, 117], "tritonserv": [114, 115, 117], "spin": [114, 115, 117], "proceed": [114, 115, 117], "flesh": [114, 115, 117], "img1": [114, 115, 117], "hakaimagazin": [114, 115, 117], "wp": [114, 115, 117], "gulf": [114, 115, 117], "bird": [114, 115, 117], "attrdict": [114, 115, 117], "pyindex": [114, 115, 117], "tritoncli": [114, 115, 117], "jump": [114, 115, 117], "firstli": [114, 115, 117], "resiz": [114, 115, 117], "httpclient": [114, 115, 117], "triton_to_np_dtyp": [114, 115, 117], "rn50_preprocess": [114, 115, 117], "img_path": [114, 115, 117], "img": [114, 115, 117], "centercrop": [114, 115, 117], "485": [114, 115, 117], "456": [114, 115, 117], "406": [114, 115, 117], "229": [114, 115, 117], "transformed_img": [114, 115, 117], "inferenceservercli": [114, 115, 117], "localhost": [114, 115, 117], "secondli": [114, 115, 117], "obtain": [114, 115, 116, 117, 121], "inferinput": [114, 115, 117], "set_data_from_numpi": [114, 115, 117], "binary_data": [114, 115, 117], "inferrequestedoutput": [114, 115, 117], "class_count": [114, 115, 117], "lastli": [114, 115, 117], "send": [114, 115, 117], "model_nam": [114, 115, 117], "inference_output": [114, 115, 117], "as_numpi": [114, 115, 117], "468750": [114, 115, 117], "90": [114, 115, 117], "523438": [114, 115, 117], "92": [114, 115, 117], "664062": [114, 115, 117], "429688": [114, 115, 117], "136": [114, 115, 117], "234375": [114, 115, 117], "confidence_scor": [114, 115, 117], "classification_index": [114, 115, 117], "_rendered_examples_python": 114, "_rendered_examples_jupyt": 114, "acoust": 116, "speech": 116, "quartznet": 116, "contextnet": 116, "subword": 116, "piec": 116, "excit": 116, "se": 116, "audio": 116, "transcrib": 116, "speedup": 116, "feedforward": 116, "cnn": 116, "uniformli": 116, "compound": 116, "coeffici": 116, "b0": 116, "english": 116, "supervis": 116, "walkthrough": 116, "adopt": 116, "mobilenetv2": 116, "classif": 116, "imagenet": 116, "imagenett": 116, "qat": 116, "simul": 116, "eagerli": 118, "swap": 118, "exactli": 118, "_tracer": 118, "queri": 118, "attn_weight": 118, "compiler_dynamic_shap": 118, "inputs_bs2": 118, "mymodul": 119, "linear1": 119, "linear2": 119, "linear3": 119, "40": 119, "__myl_mulsum_myl0_0": 119, "layertyp": 119, "kgen": 119, "__mye116_dconst": 119, "__myln_k_arg__bb1_2": 119, "tacticnam": 119, "__myl_mulsum_0xfa6c1858aea1b13b03f90165d7149ec6": 119, "streamid": 119, "__myl_addresmulsum_myl0_1": 119, "__mye131_dconst": 119, "addmm_constant_0": 119, "addmm_add_broadcast_to_same_shape_lhs_broadcast_constantfloat": 119, "__myln_k_arg__bb1_3": 119, "__myl_addresmulsum_0xb3915d7ebfe48be45b6d49083479e12f": 119, "__myl_addresmulsumadd_myl0_2": 119, "__mye146_dconst": 119, "addmm_2_constant_0": 119, "addmm_2_add_broadcast_to_same_shape_lhs_broadcast_constantfloat": 119, "addmm_1_constant_0": 119, "addmm_1_add_broadcast_to_same_shape_lhs_broadcast_constantfloat": 119, "__myl_addresmulsumadd_0xcdd0085ad25f5f45ac5fafb72acbffd6": 119, "__myl_mulsumaddcas_myl0_0": 119, "__mye112_dconst": 119, "__myl_mulsumaddcas_0xacf8f5dd9be2f3e7bb09cdddeac6c936": 119, "__myl_resmulsumaddcas_myl0_1": 119, "__mye127_dconst": 119, "addmm_1_add_broadcast_to_same_shape_lhs_broadcast_constanthalf": 119, "__myl_resmulsumaddcas_0x5a3b318b5a1c97b7d5110c0291481337": 119, "__myl_resmulsumadd_myl0_2": 119, "__mye142_dconst": 119, "__myl_resmulsumadd_0x3fad91127c640fd6db771aa9cde67db0": 119, "libtorchtrt_runtim": 120, "dl_open": 120, "ld_preload": 120, "load_librari": 120, "wl": 120, "ltorchtrt": 120, "torchtrt_runtime_exampl": 120, "libtorchtrt_plugin": 120, "neglig": 120, "alert": 120, "switch": 120, "mismatch": 120, "crash": 120, "sacrif": 120, "incur": 120, "intens": 120, "trt_ep": 121, "stai": 121, "trt_t": 121, "ergonom": 122, "deleg": 122, "believ": 122, "amen": 122, "artifact": 122, "pack": 122, "year": 122, "superset": 122, "codebas": 122, "immedi": 122, "traceabl": 122, "scriptabl": 122, "neural": 123, "deconvolut": 123, "scripted_model": 123}, "objects": {"": [[5, 0, 1, "c.STR", "STR"], [9, 0, 1, "c.TORCHTRT_API", "TORCHTRT_API"], [11, 0, 1, "c.TORCHTRT_HIDDEN", "TORCHTRT_HIDDEN"], [7, 0, 1, "c.TORCH_TENSORRT_MAJOR_VERSION", "TORCH_TENSORRT_MAJOR_VERSION"], [8, 0, 1, "c.TORCH_TENSORRT_MINOR_VERSION", "TORCH_TENSORRT_MINOR_VERSION"], [6, 0, 1, "c.TORCH_TENSORRT_PATCH_VERSION", "TORCH_TENSORRT_PATCH_VERSION"], [12, 0, 1, "c.TORCH_TENSORRT_VERSION", "TORCH_TENSORRT_VERSION"], [10, 0, 1, "c.XSTR", "XSTR"], [0, 1, 1, "_CPPv4N14torch_tensorrt8DataTypeE", "torch_tensorrt::DataType"], [0, 2, 1, "_CPPv4N14torch_tensorrt8DataType8DataTypeE5Value", "torch_tensorrt::DataType::DataType"], [0, 2, 1, "_CPPv4N14torch_tensorrt8DataType8DataTypeEN3c1010ScalarTypeE", "torch_tensorrt::DataType::DataType"], [0, 2, 1, "_CPPv4N14torch_tensorrt8DataType8DataTypeEv", "torch_tensorrt::DataType::DataType"], [0, 3, 1, "_CPPv4N14torch_tensorrt8DataType8DataTypeE5Value", "torch_tensorrt::DataType::DataType::t"], [0, 3, 1, "_CPPv4N14torch_tensorrt8DataType8DataTypeEN3c1010ScalarTypeE", "torch_tensorrt::DataType::DataType::t"], [0, 4, 1, "_CPPv4N14torch_tensorrt8DataType5ValueE", "torch_tensorrt::DataType::Value"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value5kBoolE", "torch_tensorrt::DataType::Value::kBool"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value5kCharE", "torch_tensorrt::DataType::Value::kChar"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value7kDoubleE", "torch_tensorrt::DataType::Value::kDouble"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value6kFloatE", "torch_tensorrt::DataType::Value::kFloat"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value5kHalfE", "torch_tensorrt::DataType::Value::kHalf"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value4kIntE", "torch_tensorrt::DataType::Value::kInt"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value5kLongE", "torch_tensorrt::DataType::Value::kLong"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value8kUnknownE", "torch_tensorrt::DataType::Value::kUnknown"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value5kBoolE", "torch_tensorrt::DataType::kBool"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value5kCharE", "torch_tensorrt::DataType::kChar"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value7kDoubleE", "torch_tensorrt::DataType::kDouble"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value6kFloatE", "torch_tensorrt::DataType::kFloat"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value5kHalfE", "torch_tensorrt::DataType::kHalf"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value4kIntE", "torch_tensorrt::DataType::kInt"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value5kLongE", "torch_tensorrt::DataType::kLong"], [0, 5, 1, "_CPPv4N14torch_tensorrt8DataType5Value8kUnknownE", "torch_tensorrt::DataType::kUnknown"], [0, 2, 1, "_CPPv4NK14torch_tensorrt8DataTypecv5ValueEv", "torch_tensorrt::DataType::operator Value"], [0, 2, 1, "_CPPv4N14torch_tensorrt8DataTypecvbEv", "torch_tensorrt::DataType::operator bool"], [0, 2, 1, "_CPPv4NK14torch_tensorrt8DataTypeneE8DataType", "torch_tensorrt::DataType::operator!="], [0, 2, 1, "_CPPv4NK14torch_tensorrt8DataTypeneEN8DataType5ValueE", "torch_tensorrt::DataType::operator!="], [0, 3, 1, "_CPPv4NK14torch_tensorrt8DataTypeneE8DataType", "torch_tensorrt::DataType::operator!=::other"], [0, 3, 1, "_CPPv4NK14torch_tensorrt8DataTypeneEN8DataType5ValueE", "torch_tensorrt::DataType::operator!=::other"], [0, 2, 1, "_CPPv4NK14torch_tensorrt8DataTypeeqE8DataType", "torch_tensorrt::DataType::operator=="], [0, 2, 1, "_CPPv4NK14torch_tensorrt8DataTypeeqEN8DataType5ValueE", "torch_tensorrt::DataType::operator=="], [0, 3, 1, "_CPPv4NK14torch_tensorrt8DataTypeeqE8DataType", "torch_tensorrt::DataType::operator==::other"], [0, 3, 1, "_CPPv4NK14torch_tensorrt8DataTypeeqEN8DataType5ValueE", "torch_tensorrt::DataType::operator==::other"], [46, 1, 1, "_CPPv4N14torch_tensorrt6DeviceE", "torch_tensorrt::Device"], [46, 2, 1, "_CPPv4N14torch_tensorrt6Device6DeviceEv", "torch_tensorrt::Device::Device"], [1, 1, 1, "_CPPv4N14torch_tensorrt6Device10DeviceTypeE", "torch_tensorrt::Device::DeviceType"], [46, 1, 1, "_CPPv4N14torch_tensorrt6Device10DeviceTypeE", "torch_tensorrt::Device::DeviceType"], [1, 2, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeE5Value", "torch_tensorrt::Device::DeviceType::DeviceType"], [1, 2, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeEN3c1010DeviceTypeE", "torch_tensorrt::Device::DeviceType::DeviceType"], [1, 2, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeEv", "torch_tensorrt::Device::DeviceType::DeviceType"], [46, 2, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeE5Value", "torch_tensorrt::Device::DeviceType::DeviceType"], [46, 2, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeEN3c1010DeviceTypeE", "torch_tensorrt::Device::DeviceType::DeviceType"], [46, 2, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeEv", "torch_tensorrt::Device::DeviceType::DeviceType"], [1, 3, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeE5Value", "torch_tensorrt::Device::DeviceType::DeviceType::t"], [1, 3, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeEN3c1010DeviceTypeE", "torch_tensorrt::Device::DeviceType::DeviceType::t"], [46, 3, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeE5Value", "torch_tensorrt::Device::DeviceType::DeviceType::t"], [46, 3, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeEN3c1010DeviceTypeE", "torch_tensorrt::Device::DeviceType::DeviceType::t"], [1, 4, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType5ValueE", "torch_tensorrt::Device::DeviceType::Value"], [46, 4, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType5ValueE", "torch_tensorrt::Device::DeviceType::Value"], [1, 5, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType5Value4kDLAE", "torch_tensorrt::Device::DeviceType::Value::kDLA"], [46, 5, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType5Value4kDLAE", "torch_tensorrt::Device::DeviceType::Value::kDLA"], [1, 5, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType5Value4kGPUE", "torch_tensorrt::Device::DeviceType::Value::kGPU"], [46, 5, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType5Value4kGPUE", "torch_tensorrt::Device::DeviceType::Value::kGPU"], [1, 5, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType5Value4kDLAE", "torch_tensorrt::Device::DeviceType::kDLA"], [1, 5, 1, "_CPPv4N14torch_tensorrt6Device10DeviceType5Value4kGPUE", "torch_tensorrt::Device::DeviceType::kGPU"], [1, 2, 1, "_CPPv4NK14torch_tensorrt6Device10DeviceTypecv5ValueEv", "torch_tensorrt::Device::DeviceType::operator Value"], [46, 2, 1, "_CPPv4NK14torch_tensorrt6Device10DeviceTypecv5ValueEv", "torch_tensorrt::Device::DeviceType::operator Value"], [1, 2, 1, "_CPPv4N14torch_tensorrt6Device10DeviceTypecvbEv", "torch_tensorrt::Device::DeviceType::operator bool"], [46, 2, 1, "_CPPv4N14torch_tensorrt6Device10DeviceTypecvbEv", "torch_tensorrt::Device::DeviceType::operator bool"], [1, 2, 1, "_CPPv4NK14torch_tensorrt6Device10DeviceTypeneE10DeviceType", "torch_tensorrt::Device::DeviceType::operator!="], [46, 2, 1, "_CPPv4NK14torch_tensorrt6Device10DeviceTypeneE10DeviceType", "torch_tensorrt::Device::DeviceType::operator!="], [1, 3, 1, "_CPPv4NK14torch_tensorrt6Device10DeviceTypeneE10DeviceType", "torch_tensorrt::Device::DeviceType::operator!=::other"], [46, 3, 1, "_CPPv4NK14torch_tensorrt6Device10DeviceTypeneE10DeviceType", "torch_tensorrt::Device::DeviceType::operator!=::other"], [1, 2, 1, "_CPPv4NK14torch_tensorrt6Device10DeviceTypeeqE10DeviceType", "torch_tensorrt::Device::DeviceType::operator=="], [46, 2, 1, "_CPPv4NK14torch_tensorrt6Device10DeviceTypeeqE10DeviceType", "torch_tensorrt::Device::DeviceType::operator=="], [1, 3, 1, "_CPPv4NK14torch_tensorrt6Device10DeviceTypeeqE10DeviceType", "torch_tensorrt::Device::DeviceType::operator==::other"], [46, 3, 1, "_CPPv4NK14torch_tensorrt6Device10DeviceTypeeqE10DeviceType", "torch_tensorrt::Device::DeviceType::operator==::other"], [46, 6, 1, "_CPPv4N14torch_tensorrt6Device18allow_gpu_fallbackE", "torch_tensorrt::Device::allow_gpu_fallback"], [46, 6, 1, "_CPPv4N14torch_tensorrt6Device11device_typeE", "torch_tensorrt::Device::device_type"], [46, 6, 1, "_CPPv4N14torch_tensorrt6Device8dla_coreE", "torch_tensorrt::Device::dla_core"], [46, 6, 1, "_CPPv4N14torch_tensorrt6Device6gpu_idE", "torch_tensorrt::Device::gpu_id"], [17, 4, 1, "_CPPv4N14torch_tensorrt16EngineCapabilityE", "torch_tensorrt::EngineCapability"], [17, 5, 1, "_CPPv4N14torch_tensorrt16EngineCapability15kDLA_STANDALONEE", "torch_tensorrt::EngineCapability::kDLA_STANDALONE"], [17, 5, 1, "_CPPv4N14torch_tensorrt16EngineCapability7kSAFETYE", "torch_tensorrt::EngineCapability::kSAFETY"], [17, 5, 1, "_CPPv4N14torch_tensorrt16EngineCapability9kSTANDARDE", "torch_tensorrt::EngineCapability::kSTANDARD"], [47, 1, 1, "_CPPv4N14torch_tensorrt11GraphInputsE", "torch_tensorrt::GraphInputs"], [47, 6, 1, "_CPPv4N14torch_tensorrt11GraphInputs15input_signatureE", "torch_tensorrt::GraphInputs::input_signature"], [47, 6, 1, "_CPPv4N14torch_tensorrt11GraphInputs6inputsE", "torch_tensorrt::GraphInputs::inputs"], [48, 1, 1, "_CPPv4N14torch_tensorrt5InputE", "torch_tensorrt::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputEN2at6TensorE", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input"], [48, 2, 1, "_CPPv4N14torch_tensorrt5Input5InputEv", "torch_tensorrt::Input::Input"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::dtype"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::dtype"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::dtype"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::dtype"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::dtype"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::dtype"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::dtype"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::dtype"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::format"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input::max_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::max_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::max_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::max_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input::max_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::max_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::max_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::max_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input::min_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::min_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::min_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::min_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input::min_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::min_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::min_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::min_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input::opt_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::opt_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::opt_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::opt_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input::opt_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::opt_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::opt_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::opt_shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input::shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE12TensorFormat", "torch_tensorrt::Input::Input::shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataType12TensorFormat", "torch_tensorrt::Input::Input::shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::shape"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN2at6TensorE", "torch_tensorrt::Input::Input::tensor"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::tensor_domain"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::tensor_domain"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::tensor_domain"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::tensor_domain"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::tensor_domain"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::tensor_domain"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::tensor_domain"], [48, 3, 1, "_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat", "torch_tensorrt::Input::Input::tensor_domain"], [48, 6, 1, "_CPPv4N14torch_tensorrt5Input5dtypeE", "torch_tensorrt::Input::dtype"], [48, 6, 1, "_CPPv4N14torch_tensorrt5Input6formatE", "torch_tensorrt::Input::format"], [48, 6, 1, "_CPPv4N14torch_tensorrt5Input9max_shapeE", "torch_tensorrt::Input::max_shape"], [48, 6, 1, "_CPPv4N14torch_tensorrt5Input9min_shapeE", "torch_tensorrt::Input::min_shape"], [48, 6, 1, "_CPPv4N14torch_tensorrt5Input9opt_shapeE", "torch_tensorrt::Input::opt_shape"], [48, 6, 1, "_CPPv4N14torch_tensorrt5Input5shapeE", "torch_tensorrt::Input::shape"], [48, 6, 1, "_CPPv4N14torch_tensorrt5Input13tensor_domainE", "torch_tensorrt::Input::tensor_domain"], [2, 1, 1, "_CPPv4N14torch_tensorrt12TensorFormatE", "torch_tensorrt::TensorFormat"], [2, 2, 1, "_CPPv4N14torch_tensorrt12TensorFormat12TensorFormatE5Value", "torch_tensorrt::TensorFormat::TensorFormat"], [2, 2, 1, "_CPPv4N14torch_tensorrt12TensorFormat12TensorFormatEN2at12MemoryFormatE", "torch_tensorrt::TensorFormat::TensorFormat"], [2, 2, 1, "_CPPv4N14torch_tensorrt12TensorFormat12TensorFormatEv", "torch_tensorrt::TensorFormat::TensorFormat"], [2, 3, 1, "_CPPv4N14torch_tensorrt12TensorFormat12TensorFormatE5Value", "torch_tensorrt::TensorFormat::TensorFormat::t"], [2, 3, 1, "_CPPv4N14torch_tensorrt12TensorFormat12TensorFormatEN2at12MemoryFormatE", "torch_tensorrt::TensorFormat::TensorFormat::t"], [2, 4, 1, "_CPPv4N14torch_tensorrt12TensorFormat5ValueE", "torch_tensorrt::TensorFormat::Value"], [2, 5, 1, "_CPPv4N14torch_tensorrt12TensorFormat5Value13kChannelsLastE", "torch_tensorrt::TensorFormat::Value::kChannelsLast"], [2, 5, 1, "_CPPv4N14torch_tensorrt12TensorFormat5Value11kContiguousE", "torch_tensorrt::TensorFormat::Value::kContiguous"], [2, 5, 1, "_CPPv4N14torch_tensorrt12TensorFormat5Value8kUnknownE", "torch_tensorrt::TensorFormat::Value::kUnknown"], [2, 5, 1, "_CPPv4N14torch_tensorrt12TensorFormat5Value13kChannelsLastE", "torch_tensorrt::TensorFormat::kChannelsLast"], [2, 5, 1, "_CPPv4N14torch_tensorrt12TensorFormat5Value11kContiguousE", "torch_tensorrt::TensorFormat::kContiguous"], [2, 5, 1, "_CPPv4N14torch_tensorrt12TensorFormat5Value8kUnknownE", "torch_tensorrt::TensorFormat::kUnknown"], [2, 2, 1, "_CPPv4NK14torch_tensorrt12TensorFormatcv5ValueEv", "torch_tensorrt::TensorFormat::operator Value"], [2, 2, 1, "_CPPv4N14torch_tensorrt12TensorFormatcvbEv", "torch_tensorrt::TensorFormat::operator bool"], [2, 2, 1, "_CPPv4NK14torch_tensorrt12TensorFormatneE12TensorFormat", "torch_tensorrt::TensorFormat::operator!="], [2, 2, 1, "_CPPv4NK14torch_tensorrt12TensorFormatneEN12TensorFormat5ValueE", "torch_tensorrt::TensorFormat::operator!="], [2, 3, 1, "_CPPv4NK14torch_tensorrt12TensorFormatneE12TensorFormat", "torch_tensorrt::TensorFormat::operator!=::other"], [2, 3, 1, "_CPPv4NK14torch_tensorrt12TensorFormatneEN12TensorFormat5ValueE", "torch_tensorrt::TensorFormat::operator!=::other"], [2, 2, 1, "_CPPv4NK14torch_tensorrt12TensorFormateqE12TensorFormat", "torch_tensorrt::TensorFormat::operator=="], [2, 2, 1, "_CPPv4NK14torch_tensorrt12TensorFormateqEN12TensorFormat5ValueE", "torch_tensorrt::TensorFormat::operator=="], [2, 3, 1, "_CPPv4NK14torch_tensorrt12TensorFormateqE12TensorFormat", "torch_tensorrt::TensorFormat::operator==::other"], [2, 3, 1, "_CPPv4NK14torch_tensorrt12TensorFormateqEN12TensorFormat5ValueE", "torch_tensorrt::TensorFormat::operator==::other"], [36, 2, 1, "_CPPv4N14torch_tensorrt15dump_build_infoEv", "torch_tensorrt::dump_build_info"], [34, 2, 1, "_CPPv4N14torch_tensorrt14get_build_infoEv", "torch_tensorrt::get_build_info"], [16, 4, 1, "_CPPv4N14torch_tensorrt7logging5LevelE", "torch_tensorrt::logging::Level"], [16, 5, 1, "_CPPv4N14torch_tensorrt7logging5Level6kDEBUGE", "torch_tensorrt::logging::Level::kDEBUG"], [16, 5, 1, "_CPPv4N14torch_tensorrt7logging5Level6kERRORE", "torch_tensorrt::logging::Level::kERROR"], [16, 5, 1, "_CPPv4N14torch_tensorrt7logging5Level6kGRAPHE", "torch_tensorrt::logging::Level::kGRAPH"], [16, 5, 1, "_CPPv4N14torch_tensorrt7logging5Level5kINFOE", "torch_tensorrt::logging::Level::kINFO"], [16, 5, 1, "_CPPv4N14torch_tensorrt7logging5Level15kINTERNAL_ERRORE", "torch_tensorrt::logging::Level::kINTERNAL_ERROR"], [16, 5, 1, "_CPPv4N14torch_tensorrt7logging5Level8kWARNINGE", "torch_tensorrt::logging::Level::kWARNING"], [24, 2, 1, "_CPPv4N14torch_tensorrt7logging24get_is_colored_output_onEv", "torch_tensorrt::logging::get_is_colored_output_on"], [22, 2, 1, "_CPPv4N14torch_tensorrt7logging18get_logging_prefixEv", "torch_tensorrt::logging::get_logging_prefix"], [23, 2, 1, "_CPPv4N14torch_tensorrt7logging24get_reportable_log_levelEv", "torch_tensorrt::logging::get_reportable_log_level"], [16, 5, 1, "_CPPv4N14torch_tensorrt7logging5Level6kDEBUGE", "torch_tensorrt::logging::kDEBUG"], [16, 5, 1, "_CPPv4N14torch_tensorrt7logging5Level6kERRORE", "torch_tensorrt::logging::kERROR"], [16, 5, 1, "_CPPv4N14torch_tensorrt7logging5Level6kGRAPHE", "torch_tensorrt::logging::kGRAPH"], [16, 5, 1, "_CPPv4N14torch_tensorrt7logging5Level5kINFOE", "torch_tensorrt::logging::kINFO"], [16, 5, 1, "_CPPv4N14torch_tensorrt7logging5Level15kINTERNAL_ERRORE", "torch_tensorrt::logging::kINTERNAL_ERROR"], [16, 5, 1, "_CPPv4N14torch_tensorrt7logging5Level8kWARNINGE", "torch_tensorrt::logging::kWARNING"], [26, 2, 1, "_CPPv4N14torch_tensorrt7logging3logE5LevelNSt6stringE", "torch_tensorrt::logging::log"], [26, 3, 1, "_CPPv4N14torch_tensorrt7logging3logE5LevelNSt6stringE", "torch_tensorrt::logging::log::lvl"], [26, 3, 1, "_CPPv4N14torch_tensorrt7logging3logE5LevelNSt6stringE", "torch_tensorrt::logging::log::msg"], [27, 2, 1, "_CPPv4N14torch_tensorrt7logging24set_is_colored_output_onEb", "torch_tensorrt::logging::set_is_colored_output_on"], [27, 3, 1, "_CPPv4N14torch_tensorrt7logging24set_is_colored_output_onEb", "torch_tensorrt::logging::set_is_colored_output_on::colored_output_on"], [28, 2, 1, "_CPPv4N14torch_tensorrt7logging18set_logging_prefixENSt6stringE", "torch_tensorrt::logging::set_logging_prefix"], [28, 3, 1, "_CPPv4N14torch_tensorrt7logging18set_logging_prefixENSt6stringE", "torch_tensorrt::logging::set_logging_prefix::prefix"], [25, 2, 1, "_CPPv4N14torch_tensorrt7logging24set_reportable_log_levelE5Level", "torch_tensorrt::logging::set_reportable_log_level"], [25, 3, 1, "_CPPv4N14torch_tensorrt7logging24set_reportable_log_levelE5Level", "torch_tensorrt::logging::set_reportable_log_level::lvl"], [3, 1, 1, "_CPPv4I0EN14torch_tensorrt3ptq19Int8CacheCalibratorE", "torch_tensorrt::ptq::Int8CacheCalibrator"], [3, 7, 1, "_CPPv4I0EN14torch_tensorrt3ptq19Int8CacheCalibratorE", "torch_tensorrt::ptq::Int8CacheCalibrator::Algorithm"], [3, 2, 1, "_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibrator19Int8CacheCalibratorERKNSt6stringE", "torch_tensorrt::ptq::Int8CacheCalibrator::Int8CacheCalibrator"], [3, 3, 1, "_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibrator19Int8CacheCalibratorERKNSt6stringE", "torch_tensorrt::ptq::Int8CacheCalibrator::Int8CacheCalibrator::cache_file_path"], [3, 2, 1, "_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibrator8getBatchEA_PvA_PKci", "torch_tensorrt::ptq::Int8CacheCalibrator::getBatch"], [3, 3, 1, "_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibrator8getBatchEA_PvA_PKci", "torch_tensorrt::ptq::Int8CacheCalibrator::getBatch::bindings"], [3, 3, 1, "_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibrator8getBatchEA_PvA_PKci", "torch_tensorrt::ptq::Int8CacheCalibrator::getBatch::names"], [3, 3, 1, "_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibrator8getBatchEA_PvA_PKci", "torch_tensorrt::ptq::Int8CacheCalibrator::getBatch::nbBindings"], [3, 2, 1, "_CPPv4NK14torch_tensorrt3ptq19Int8CacheCalibrator12getBatchSizeEv", "torch_tensorrt::ptq::Int8CacheCalibrator::getBatchSize"], [3, 2, 1, "_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibratorcvPN8nvinfer115IInt8CalibratorEEv", "torch_tensorrt::ptq::Int8CacheCalibrator::operator nvinfer1::IInt8Calibrator*"], [3, 2, 1, "_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibrator20readCalibrationCacheER6size_t", "torch_tensorrt::ptq::Int8CacheCalibrator::readCalibrationCache"], [3, 3, 1, "_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibrator20readCalibrationCacheER6size_t", "torch_tensorrt::ptq::Int8CacheCalibrator::readCalibrationCache::length"], [3, 2, 1, "_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibrator21writeCalibrationCacheEPKv6size_t", "torch_tensorrt::ptq::Int8CacheCalibrator::writeCalibrationCache"], [3, 3, 1, "_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibrator21writeCalibrationCacheEPKv6size_t", "torch_tensorrt::ptq::Int8CacheCalibrator::writeCalibrationCache::cache"], [3, 3, 1, "_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibrator21writeCalibrationCacheEPKv6size_t", "torch_tensorrt::ptq::Int8CacheCalibrator::writeCalibrationCache::length"], [4, 1, 1, "_CPPv4I00EN14torch_tensorrt3ptq14Int8CalibratorE", "torch_tensorrt::ptq::Int8Calibrator"], [4, 7, 1, "_CPPv4I00EN14torch_tensorrt3ptq14Int8CalibratorE", "torch_tensorrt::ptq::Int8Calibrator::Algorithm"], [4, 7, 1, "_CPPv4I00EN14torch_tensorrt3ptq14Int8CalibratorE", "torch_tensorrt::ptq::Int8Calibrator::DataLoaderUniquePtr"], [4, 2, 1, "_CPPv4N14torch_tensorrt3ptq14Int8Calibrator14Int8CalibratorE19DataLoaderUniquePtrRKNSt6stringEb", "torch_tensorrt::ptq::Int8Calibrator::Int8Calibrator"], [4, 3, 1, "_CPPv4N14torch_tensorrt3ptq14Int8Calibrator14Int8CalibratorE19DataLoaderUniquePtrRKNSt6stringEb", "torch_tensorrt::ptq::Int8Calibrator::Int8Calibrator::cache_file_path"], [4, 3, 1, "_CPPv4N14torch_tensorrt3ptq14Int8Calibrator14Int8CalibratorE19DataLoaderUniquePtrRKNSt6stringEb", "torch_tensorrt::ptq::Int8Calibrator::Int8Calibrator::dataloader"], [4, 3, 1, "_CPPv4N14torch_tensorrt3ptq14Int8Calibrator14Int8CalibratorE19DataLoaderUniquePtrRKNSt6stringEb", "torch_tensorrt::ptq::Int8Calibrator::Int8Calibrator::use_cache"], [4, 2, 1, "_CPPv4N14torch_tensorrt3ptq14Int8Calibrator8getBatchEA_PvA_PKci", "torch_tensorrt::ptq::Int8Calibrator::getBatch"], [4, 3, 1, "_CPPv4N14torch_tensorrt3ptq14Int8Calibrator8getBatchEA_PvA_PKci", "torch_tensorrt::ptq::Int8Calibrator::getBatch::bindings"], [4, 3, 1, "_CPPv4N14torch_tensorrt3ptq14Int8Calibrator8getBatchEA_PvA_PKci", "torch_tensorrt::ptq::Int8Calibrator::getBatch::names"], [4, 3, 1, "_CPPv4N14torch_tensorrt3ptq14Int8Calibrator8getBatchEA_PvA_PKci", "torch_tensorrt::ptq::Int8Calibrator::getBatch::nbBindings"], [4, 2, 1, "_CPPv4NK14torch_tensorrt3ptq14Int8Calibrator12getBatchSizeEv", "torch_tensorrt::ptq::Int8Calibrator::getBatchSize"], [4, 2, 1, "_CPPv4N14torch_tensorrt3ptq14Int8CalibratorcvPN8nvinfer115IInt8CalibratorEEv", "torch_tensorrt::ptq::Int8Calibrator::operator nvinfer1::IInt8Calibrator*"], [4, 2, 1, "_CPPv4N14torch_tensorrt3ptq14Int8Calibrator20readCalibrationCacheER6size_t", "torch_tensorrt::ptq::Int8Calibrator::readCalibrationCache"], [4, 3, 1, "_CPPv4N14torch_tensorrt3ptq14Int8Calibrator20readCalibrationCacheER6size_t", "torch_tensorrt::ptq::Int8Calibrator::readCalibrationCache::length"], [4, 2, 1, "_CPPv4N14torch_tensorrt3ptq14Int8Calibrator21writeCalibrationCacheEPKv6size_t", "torch_tensorrt::ptq::Int8Calibrator::writeCalibrationCache"], [4, 3, 1, "_CPPv4N14torch_tensorrt3ptq14Int8Calibrator21writeCalibrationCacheEPKv6size_t", "torch_tensorrt::ptq::Int8Calibrator::writeCalibrationCache::cache"], [4, 3, 1, "_CPPv4N14torch_tensorrt3ptq14Int8Calibrator21writeCalibrationCacheEPKv6size_t", "torch_tensorrt::ptq::Int8Calibrator::writeCalibrationCache::length"], [29, 2, 1, "_CPPv4I0EN14torch_tensorrt3ptq26make_int8_cache_calibratorE19Int8CacheCalibratorI9AlgorithmERKNSt6stringE", "torch_tensorrt::ptq::make_int8_cache_calibrator"], [29, 7, 1, "_CPPv4I0EN14torch_tensorrt3ptq26make_int8_cache_calibratorE19Int8CacheCalibratorI9AlgorithmERKNSt6stringE", "torch_tensorrt::ptq::make_int8_cache_calibrator::Algorithm"], [29, 3, 1, "_CPPv4I0EN14torch_tensorrt3ptq26make_int8_cache_calibratorE19Int8CacheCalibratorI9AlgorithmERKNSt6stringE", "torch_tensorrt::ptq::make_int8_cache_calibrator::cache_file_path"], [30, 2, 1, "_CPPv4I00EN14torch_tensorrt3ptq20make_int8_calibratorE14Int8CalibratorI9Algorithm10DataLoaderE10DataLoaderRKNSt6stringEb", "torch_tensorrt::ptq::make_int8_calibrator"], [30, 7, 1, "_CPPv4I00EN14torch_tensorrt3ptq20make_int8_calibratorE14Int8CalibratorI9Algorithm10DataLoaderE10DataLoaderRKNSt6stringEb", "torch_tensorrt::ptq::make_int8_calibrator::Algorithm"], [30, 7, 1, "_CPPv4I00EN14torch_tensorrt3ptq20make_int8_calibratorE14Int8CalibratorI9Algorithm10DataLoaderE10DataLoaderRKNSt6stringEb", "torch_tensorrt::ptq::make_int8_calibrator::DataLoader"], [30, 3, 1, "_CPPv4I00EN14torch_tensorrt3ptq20make_int8_calibratorE14Int8CalibratorI9Algorithm10DataLoaderE10DataLoaderRKNSt6stringEb", "torch_tensorrt::ptq::make_int8_calibrator::cache_file_path"], [30, 3, 1, "_CPPv4I00EN14torch_tensorrt3ptq20make_int8_calibratorE14Int8CalibratorI9Algorithm10DataLoaderE10DataLoaderRKNSt6stringEb", "torch_tensorrt::ptq::make_int8_calibrator::dataloader"], [30, 3, 1, "_CPPv4I00EN14torch_tensorrt3ptq20make_int8_calibratorE14Int8CalibratorI9Algorithm10DataLoaderE10DataLoaderRKNSt6stringEb", "torch_tensorrt::ptq::make_int8_calibrator::use_cache"], [35, 2, 1, "_CPPv4N14torch_tensorrt10set_deviceEKi", "torch_tensorrt::set_device"], [35, 3, 1, "_CPPv4N14torch_tensorrt10set_deviceEKi", "torch_tensorrt::set_device::gpu_id"], [49, 1, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpecE", "torch_tensorrt::torchscript::CompileSpec"], [49, 2, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecEN5torch3jit6IValueE", "torch_tensorrt::torchscript::CompileSpec::CompileSpec"], [49, 2, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecENSt6vectorI5InputEE", "torch_tensorrt::torchscript::CompileSpec::CompileSpec"], [49, 2, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecENSt6vectorIN3c108ArrayRefI7int64_tEEEE", "torch_tensorrt::torchscript::CompileSpec::CompileSpec"], [49, 2, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecENSt6vectorINSt6vectorI7int64_tEEEE", "torch_tensorrt::torchscript::CompileSpec::CompileSpec"], [49, 3, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecENSt6vectorIN3c108ArrayRefI7int64_tEEEE", "torch_tensorrt::torchscript::CompileSpec::CompileSpec::fixed_sizes"], [49, 3, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecENSt6vectorINSt6vectorI7int64_tEEEE", "torch_tensorrt::torchscript::CompileSpec::CompileSpec::fixed_sizes"], [49, 3, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecEN5torch3jit6IValueE", "torch_tensorrt::torchscript::CompileSpec::CompileSpec::input_signature"], [49, 3, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecENSt6vectorI5InputEE", "torch_tensorrt::torchscript::CompileSpec::CompileSpec::inputs"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec19allow_shape_tensorsE", "torch_tensorrt::torchscript::CompileSpec::allow_shape_tensors"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec10capabilityE", "torch_tensorrt::torchscript::CompileSpec::capability"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec5debugE", "torch_tensorrt::torchscript::CompileSpec::debug"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec6deviceE", "torch_tensorrt::torchscript::CompileSpec::device"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec12disable_tf32E", "torch_tensorrt::torchscript::CompileSpec::disable_tf32"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec20dla_global_dram_sizeE", "torch_tensorrt::torchscript::CompileSpec::dla_global_dram_size"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec19dla_local_dram_sizeE", "torch_tensorrt::torchscript::CompileSpec::dla_local_dram_size"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec13dla_sram_sizeE", "torch_tensorrt::torchscript::CompileSpec::dla_sram_size"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec18enabled_precisionsE", "torch_tensorrt::torchscript::CompileSpec::enabled_precisions"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec12graph_inputsE", "torch_tensorrt::torchscript::CompileSpec::graph_inputs"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec14min_block_sizeE", "torch_tensorrt::torchscript::CompileSpec::min_block_size"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec20num_avg_timing_itersE", "torch_tensorrt::torchscript::CompileSpec::num_avg_timing_iters"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec14ptq_calibratorE", "torch_tensorrt::torchscript::CompileSpec::ptq_calibrator"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec5refitE", "torch_tensorrt::torchscript::CompileSpec::refit"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec24require_full_compilationE", "torch_tensorrt::torchscript::CompileSpec::require_full_compilation"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec14sparse_weightsE", "torch_tensorrt::torchscript::CompileSpec::sparse_weights"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec22torch_executed_modulesE", "torch_tensorrt::torchscript::CompileSpec::torch_executed_modules"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec18torch_executed_opsE", "torch_tensorrt::torchscript::CompileSpec::torch_executed_ops"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec24truncate_long_and_doubleE", "torch_tensorrt::torchscript::CompileSpec::truncate_long_and_double"], [49, 6, 1, "_CPPv4N14torch_tensorrt11torchscript11CompileSpec14workspace_sizeE", "torch_tensorrt::torchscript::CompileSpec::workspace_size"], [31, 2, 1, "_CPPv4N14torch_tensorrt11torchscript29check_method_operator_supportERKN5torch3jit6ModuleENSt6stringE", "torch_tensorrt::torchscript::check_method_operator_support"], [31, 3, 1, "_CPPv4N14torch_tensorrt11torchscript29check_method_operator_supportERKN5torch3jit6ModuleENSt6stringE", "torch_tensorrt::torchscript::check_method_operator_support::method_name"], [31, 3, 1, "_CPPv4N14torch_tensorrt11torchscript29check_method_operator_supportERKN5torch3jit6ModuleENSt6stringE", "torch_tensorrt::torchscript::check_method_operator_support::module"], [32, 2, 1, "_CPPv4N14torch_tensorrt11torchscript7compileERKN5torch3jit6ModuleE11CompileSpec", "torch_tensorrt::torchscript::compile"], [32, 3, 1, "_CPPv4N14torch_tensorrt11torchscript7compileERKN5torch3jit6ModuleE11CompileSpec", "torch_tensorrt::torchscript::compile::info"], [32, 3, 1, "_CPPv4N14torch_tensorrt11torchscript7compileERKN5torch3jit6ModuleE11CompileSpec", "torch_tensorrt::torchscript::compile::module"], [37, 2, 1, "_CPPv4N14torch_tensorrt11torchscript28convert_method_to_trt_engineERKN5torch3jit6ModuleENSt6stringE11CompileSpec", "torch_tensorrt::torchscript::convert_method_to_trt_engine"], [37, 3, 1, "_CPPv4N14torch_tensorrt11torchscript28convert_method_to_trt_engineERKN5torch3jit6ModuleENSt6stringE11CompileSpec", "torch_tensorrt::torchscript::convert_method_to_trt_engine::info"], [37, 3, 1, "_CPPv4N14torch_tensorrt11torchscript28convert_method_to_trt_engineERKN5torch3jit6ModuleENSt6stringE11CompileSpec", "torch_tensorrt::torchscript::convert_method_to_trt_engine::method_name"], [37, 3, 1, "_CPPv4N14torch_tensorrt11torchscript28convert_method_to_trt_engineERKN5torch3jit6ModuleENSt6stringE11CompileSpec", "torch_tensorrt::torchscript::convert_method_to_trt_engine::module"], [33, 2, 1, "_CPPv4N14torch_tensorrt11torchscript26embed_engine_in_new_moduleERKNSt6stringE6DeviceRKNSt6vectorINSt6stringEEERKNSt6vectorINSt6stringEEE", "torch_tensorrt::torchscript::embed_engine_in_new_module"], [33, 3, 1, "_CPPv4N14torch_tensorrt11torchscript26embed_engine_in_new_moduleERKNSt6stringE6DeviceRKNSt6vectorINSt6stringEEERKNSt6vectorINSt6stringEEE", "torch_tensorrt::torchscript::embed_engine_in_new_module::device"], [33, 3, 1, "_CPPv4N14torch_tensorrt11torchscript26embed_engine_in_new_moduleERKNSt6stringE6DeviceRKNSt6vectorINSt6stringEEERKNSt6vectorINSt6stringEEE", "torch_tensorrt::torchscript::embed_engine_in_new_module::engine"], [33, 3, 1, "_CPPv4N14torch_tensorrt11torchscript26embed_engine_in_new_moduleERKNSt6stringE6DeviceRKNSt6vectorINSt6stringEEERKNSt6vectorINSt6stringEEE", "torch_tensorrt::torchscript::embed_engine_in_new_module::input_binding_names"], [33, 3, 1, "_CPPv4N14torch_tensorrt11torchscript26embed_engine_in_new_moduleERKNSt6stringE6DeviceRKNSt6vectorINSt6stringEEERKNSt6vectorINSt6stringEEE", "torch_tensorrt::torchscript::embed_engine_in_new_module::output_binding_names"], [76, 8, 0, "-", "torch_tensorrt"]], "torch_tensorrt": [[76, 9, 1, "", "Device"], [76, 9, 1, "", "DeviceType"], [76, 9, 1, "", "EngineCapability"], [76, 9, 1, "", "Input"], [76, 9, 1, "", "MutableTorchTensorRTModule"], [76, 12, 1, "", "compile"], [76, 12, 1, "", "convert_method_to_trt_engine"], [76, 9, 1, "", "dtype"], [121, 8, 0, "-", "dynamo"], [72, 8, 0, "-", "fx"], [76, 12, 1, "", "load"], [73, 8, 0, "-", "logging"], [76, 9, 1, "", "memory_format"], [75, 8, 0, "-", "runtime"], [76, 12, 1, "", "save"], [77, 8, 0, "-", "ts"]], "torch_tensorrt.Device": [[76, 10, 1, "", "__init__"], [76, 11, 1, "", "device_type"], [76, 11, 1, "", "dla_core"], [76, 11, 1, "", "gpu_id"]], "torch_tensorrt.DeviceType": [[76, 11, 1, "", "DLA"], [76, 11, 1, "", "GPU"], [76, 11, 1, "", "UNKNOWN"], [76, 10, 1, "", "to"], [76, 10, 1, "", "try_from"], [76, 10, 1, "", "try_to"]], "torch_tensorrt.EngineCapability": [[76, 11, 1, "", "DLA_STANDALONE"], [76, 11, 1, "", "SAFETY"], [76, 11, 1, "", "STANDARD"], [76, 10, 1, "", "to"], [76, 10, 1, "", "try_from"], [76, 10, 1, "", "try_to"]], "torch_tensorrt.Input": [[76, 10, 1, "", "__init__"], [76, 11, 1, "", "dtype"], [76, 10, 1, "", "example_tensor"], [76, 11, 1, "", "format"], [76, 10, 1, "", "from_tensor"], [76, 10, 1, "", "from_tensors"]], "torch_tensorrt.MutableTorchTensorRTModule": [[76, 10, 1, "", "__init__"], [76, 10, 1, "", "compile"], [76, 10, 1, "", "refit_gm"]], "torch_tensorrt.dtype": [[76, 11, 1, "", "b"], [76, 11, 1, "", "bf16"], [76, 11, 1, "", "f16"], [76, 11, 1, "", "f32"], [76, 11, 1, "", "f64"], [76, 11, 1, "", "f8"], [76, 11, 1, "", "i32"], [76, 11, 1, "", "i64"], [76, 11, 1, "", "i8"], [76, 10, 1, "", "to"], [76, 10, 1, "", "try_from"], [76, 10, 1, "", "try_to"], [76, 11, 1, "", "u8"], [76, 11, 1, "", "unknown"]], "torch_tensorrt.dynamo": [[71, 9, 1, "", "CompilationSettings"], [71, 12, 1, "", "compile"], [71, 12, 1, "", "export"], [71, 12, 1, "", "refit_module_weights"], [71, 12, 1, "", "trace"]], "torch_tensorrt.fx": [[72, 9, 1, "", "InputTensorSpec"], [72, 9, 1, "", "TRTInterpreter"], [72, 9, 1, "", "TRTInterpreterResult"], [72, 9, 1, "", "TRTModule"], [72, 12, 1, "", "compile"]], "torch_tensorrt.logging": [[73, 9, 1, "", "debug"], [73, 9, 1, "", "errors"], [73, 9, 1, "", "graphs"], [73, 9, 1, "", "info"], [73, 9, 1, "", "internal_errors"], [73, 9, 1, "", "warnings"]], "torch_tensorrt.memory_format": [[76, 11, 1, "", "cdhw32"], [76, 11, 1, "", "chw16"], [76, 11, 1, "", "chw2"], [76, 11, 1, "", "chw32"], [76, 11, 1, "", "chw4"], [76, 11, 1, "", "dhwc"], [76, 11, 1, "", "dhwc8"], [76, 11, 1, "", "dla_hwc4"], [76, 11, 1, "", "dla_linear"], [76, 11, 1, "", "hwc"], [76, 11, 1, "", "hwc16"], [76, 11, 1, "", "hwc8"], [76, 11, 1, "", "linear"], [76, 10, 1, "", "to"], [76, 10, 1, "", "try_from"], [76, 10, 1, "", "try_to"]], "torch_tensorrt.runtime": [[75, 9, 1, "", "PythonTorchTensorRTModule"], [75, 9, 1, "", "TorchTensorRTModule"], [75, 12, 1, "", "set_multi_device_safe_mode"]], "torch_tensorrt.runtime.PythonTorchTensorRTModule": [[75, 10, 1, "", "__init__"], [75, 10, 1, "", "disable_profiling"], [75, 10, 1, "", "enable_profiling"], [75, 10, 1, "", "forward"], [75, 10, 1, "", "get_layer_info"], [75, 10, 1, "", "validate_input_shapes"]], "torch_tensorrt.runtime.TorchTensorRTModule": [[75, 10, 1, "", "__init__"], [75, 10, 1, "", "forward"], [75, 10, 1, "", "get_extra_state"], [75, 10, 1, "", "set_extra_state"]], "torch_tensorrt.ts": [[77, 12, 1, "", "TensorRTCompileSpec"], [77, 12, 1, "", "check_method_op_support"], [77, 12, 1, "", "compile"], [77, 12, 1, "", "convert_method_to_trt_engine"], [77, 12, 1, "", "embed_engine_in_new_module"], [74, 8, 0, "-", "ptq"]], "torch_tensorrt.ts.ptq": [[74, 9, 1, "", "CacheCalibrator"], [74, 9, 1, "", "CalibrationAlgo"], [74, 9, 1, "", "DataLoaderCalibrator"]], "torch_tensorrt.ts.ptq.CalibrationAlgo": [[74, 11, 1, "", "ENTROPY_CALIBRATION"], [74, 11, 1, "", "ENTROPY_CALIBRATION_2"], [74, 11, 1, "", "LEGACY_CALIBRATION"], [74, 11, 1, "", "MINMAX_CALIBRATION"]]}, "objtypes": {"0": "c:macro", "1": "cpp:class", "2": "cpp:function", "3": "cpp:functionParam", "4": "cpp:enum", "5": "cpp:enumerator", "6": "cpp:member", "7": "cpp:templateParam", "8": "py:module", "9": "py:class", "10": "py:method", "11": "py:attribute", "12": "py:function"}, "objnames": {"0": ["c", "macro", "C macro"], "1": ["cpp", "class", "C++ class"], "2": ["cpp", "function", "C++ function"], "3": ["cpp", "functionParam", "C++ function parameter"], "4": ["cpp", "enum", "C++ enum"], "5": ["cpp", "enumerator", "C++ enumerator"], "6": ["cpp", "member", "C++ member"], "7": ["cpp", "templateParam", "C++ template parameter"], "8": ["py", "module", "Python module"], "9": ["py", "class", "Python class"], "10": ["py", "method", "Python method"], "11": ["py", "attribute", "Python attribute"], "12": ["py", "function", "Python function"]}, "titleterms": {"class": [0, 1, 2, 3, 4, 20, 21, 38, 40, 41, 50, 71, 72, 74, 75, 76], "datatyp": 0, "document": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 16, 17, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 46, 47, 48, 49, 61, 69, 85, 86], "devic": [1, 46, 120], "devicetyp": 1, "nest": [1, 46], "relationship": [1, 3, 4, 46, 48], "tensorformat": 2, "templat": [3, 4, 29, 30], "int8cachecalibr": 3, "inherit": [3, 4, 48], "base": [3, 4, 48, 80], "type": [3, 4, 46, 48, 54], "int8calibr": 4, "defin": [5, 6, 7, 8, 9, 10, 11, 12, 19, 50, 101, 104, 111, 112], "str": 5, "torch_tensorrt_patch_vers": 6, "torch_tensorrt_major_vers": 7, "torch_tensorrt_minor_vers": 8, "torchtrt_api": 9, "xstr": 10, "torchtrt_hidden": 11, "torch_tensorrt_vers": 12, "directori": [13, 14, 15, 51], "cpp": [13, 18, 19, 20, 21, 56], "subdirectori": [13, 14], "includ": [14, 18, 19, 20, 21], "torch_tensorrt": [15, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 45, 67, 71, 72, 73, 74, 75, 76, 77, 105, 107, 108, 122], "file": [15, 18, 19, 20, 21, 42, 43, 44, 45, 50, 51], "enum": [16, 17, 18, 21, 38, 39, 50, 74, 76], "level": [16, 80, 82, 83], "enginecap": 17, "log": [18, 22, 23, 24, 25, 26, 27, 28, 39, 42, 73], "h": [18, 19, 20, 21, 42, 43, 44, 45, 56], "content": [18, 19, 20, 21, 38, 39, 40, 41, 80, 81, 82, 83, 84, 85], "definit": [18, 19, 20, 21, 83, 95, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 112, 113], "By": [18, 19], "namespac": [18, 19, 20, 21, 38, 39, 40, 41, 50], "function": [18, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 50, 61, 71, 72, 75, 76, 77, 101, 112], "macro": [19, 43], "ptq": [20, 29, 30, 40, 44, 74, 91, 112], "get_logging_prefix": 22, "get_reportable_log_level": 23, "get_is_colored_output_on": 24, "set_reportable_log_level": 25, "set_is_colored_output_on": 27, "set_logging_prefix": 28, "make_int8_cache_calibr": 29, "make_int8_calibr": 30, "torchscript": [31, 32, 33, 37, 41, 60, 66, 69, 88, 89, 92, 121, 122], "check_method_operator_support": 31, "compil": [32, 57, 59, 63, 64, 66, 68, 69, 89, 95, 98, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 113, 116, 118, 119, 121, 122], "embed_engine_in_new_modul": 33, "get_build_info": 34, "set_devic": 35, "dump_build_info": 36, "convert_method_to_trt_engin": 37, "program": [42, 43, 44, 45, 63, 102, 120], "list": [42, 43, 44, 45, 83], "struct": [46, 47, 48, 49, 50], "graphinput": 47, "input": [48, 105, 107, 111], "compilespec": 49, "torch": [50, 61, 63, 64, 65, 66, 68, 69, 89, 90, 92, 94, 96, 100, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 114, 115, 116, 117, 118, 119, 120, 121, 122], "tensorrt": [50, 58, 61, 63, 64, 65, 66, 69, 89, 90, 92, 93, 94, 96, 100, 102, 104, 109, 110, 111, 112, 114, 115, 116, 117, 118, 119, 120, 121, 122], "c": [50, 61, 66, 68, 69, 89, 91, 116], "api": [50, 51, 61, 66, 69, 101], "hierarchi": 50, "full": [50, 51], "torchtrtc": [52, 89], "convers": [53, 57, 59, 60], "phase": [53, 55, 56, 57, 58, 59], "node": 53, "evalu": [53, 54, 70], "convert": [53, 54, 60, 65, 70, 89, 93, 94], "write": [54, 60, 62, 93, 94, 96], "dynamo": [54, 62, 69, 71, 109, 110, 111, 121, 122], "implement": [54, 94], "registr": 54, "capabl": 54, "valid": 54, "contract": [54, 60], "exampl": [54, 62, 82, 84, 95], "convolut": 54, "oper": [54, 64, 70, 89, 93, 96], "decomposit": 54, "addmm": [54, 55], "lower": [55, 57, 59, 62], "pass": [55, 62], "us": [55, 61, 89, 90, 92, 93, 94, 96, 101, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 116, 118], "eliminatecommonsubexpress": 55, "elimin": 55, "dead": 55, "code": [55, 69, 82], "except": 55, "Or": 55, "pattern": 55, "redund": 55, "guard": 55, "freez": 55, "modul": [55, 88, 89, 100, 108, 122], "fuse": 55, "branch": 55, "linear": 55, "flatten": 55, "graph": [55, 58, 108, 122], "tupl": 55, "fallback": [55, 56], "peephol": 55, "optim": [55, 68, 114, 115, 117], "remov": 55, "contigu": 55, "dropout": 55, "To": 55, "unpack": 55, "logsoftmax": 55, "unrol": 55, "loop": [55, 112], "replac": [55, 82], "tile": 55, "repeat": 55, "partit": [56, 57, 59], "partitoninfo": 56, "segmentedblock": 56, "shape_analysi": 56, "automat": [56, 93, 113], "depend": [56, 66, 99, 114], "awar": [56, 116], "runtim": [57, 58, 59, 75, 95, 101, 120], "background": [58, 60], "engin": [58, 65, 96, 97, 98], "executor": 58, "op": [58, 65, 96], "construct": 58, "result": 58, "serial": [58, 64, 68], "deseri": 58, "abi": [58, 66], "version": [58, 66], "format": [58, 122], "system": [59, 66, 93], "overview": [59, 67], "what": 60, "guarante": 60, "respons": 60, "context": [60, 80, 113], "arg": [60, 81], "weight": [60, 102, 111, 112, 113], "other": 60, "advic": 60, "link": [61, 82], "develop": 61, "avail": 61, "layer": 61, "expect": 61, "dimens": 61, "python": [61, 66, 68, 69, 88, 90, 91], "sometim": 61, "easier": 61, "read": 61, "pytorch": [61, 65, 69, 92, 93, 96, 104, 109, 110, 116], "native_op": 61, "ir": [61, 121, 122], "aten": 62, "basic": 62, "requir": 62, "regist": [62, 89], "export": [63, 68, 108, 118], "customiz": [63, 64], "set": [63, 64, 100, 103, 108, 114, 115, 117], "under": [63, 89, 118], "hood": [63, 89, 118], "trace": 63, "backend": [64, 105, 106, 107, 109, 110, 111], "kei": 64, "featur": [64, 101], "custom": [64, 89, 93, 94, 96, 98, 103, 118], "usag": [64, 102, 103], "after": 64, "model": [64, 65, 69, 93, 95, 96, 99, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 119, 121], "perform": [64, 101], "coverag": 64, "feasibl": 64, "dynam": [64, 105, 116, 118], "shape": [64, 105, 116, 118], "support": [64, 70], "recompil": [64, 105], "condit": 64, "fx": [65, 69, 72, 116, 122], "frontend": [65, 66, 69, 92, 104, 116, 122], "user": [65, 69], "guid": [65, 69], "acc": 65, "tracer": 65, "fx2trt": 65, "how": [65, 80, 91], "add": 65, "miss": 65, "instal": [66, 87], "precompil": 66, "binari": 66, "specif": 66, "cuda": [66, 103, 107, 108], "nightli": 66, "build": [66, 67, 80, 114, 115, 117], "onli": 66, "from": [66, 92], "sourc": 66, "linux": 66, "packag": [66, 120], "addit": 66, "option": [66, 68, 80, 81, 83, 105, 107, 113, 122], "distribut": 66, "No": 66, "librari": [66, 104, 111, 120], "standalon": 66, "releas": 66, "debug": 66, "pre": [66, 101, 112], "cxx11": 66, "choos": 66, "right": 66, "window": [66, 95], "step": [66, 68, 114, 115, 117], "advanc": [66, 102, 103], "setup": 66, "troubleshoot": 66, "altern": 66, "cmake": 66, "nativ": 66, "aarch64": 66, "jetson": 66, "prerequisit": [66, 67], "environ": 66, "cli": [66, 69], "jetpack": 67, "6": [67, 84], "1": [67, 68, 84, 114, 115, 117], "quick": [68, 93], "start": [68, 69], "2": [68, 84, 85, 114, 115, 117], "deploi": [68, 93, 112, 116, 120], "deploy": 68, "In": [69, 102], "framework": 69, "infer": [69, 101, 104, 105, 106, 107, 108, 112, 114, 115, 117], "nvidia": 69, "gpu": 69, "get": 69, "tutori": [69, 114], "zoo": [69, 99, 114], "contributor": 69, "indic": 69, "legaci": [69, 116, 122], "further": 69, "inform": 69, "current": 70, "through": 70, "ts": [74, 77, 122], "submodul": 76, "comput": 78, "time": [78, 122], "changelog": 79, "configur": 80, "project": 80, "wide": 80, "html": 80, "theme": [80, 86], "toc": 80, "page": 80, "tabl": [80, 81, 82, 83, 84, 85], "mod": 81, "test_py_modul": 81, "gener": [81, 93, 109, 110], "index": 81, "paramet": [81, 104], "data": 81, "paragraph": [82, 85], "markup": 82, "inlin": 82, "math": 82, "meta": 82, "block": 82, "liter": 82, "line": 82, "quot": 82, "doctest": 82, "emphas": 82, "number": [82, 83], "sidebar": 82, "ch": 82, "ien": 82, "The": [82, 89], "creativ": 82, "A": 82, "refer": [82, 111], "footnot": 82, "citat": [82, 91], "glossari": 82, "target": 82, "direct": 82, "center": 82, "text": 82, "imag": [82, 83, 111], "figur": 82, "admonit": 82, "And": 82, "wai": 82, "topic": 82, "rubric": 82, "titl": 82, "compound": 82, "download": [82, 87], "enumer": 83, "field": 83, "bullet": 83, "second": 83, "But": 83, "deeper": 83, "down": 83, "rabbit": 83, "hole": 83, "hlist": 83, "grid": 83, "giant": 83, "can": 83, "have": 83, "caption": [83, 86], "like": 83, "thi": [83, 86], "one": 83, "long": [84, 86], "sticki": 84, "nav": 84, "menu": [84, 86], "3": [84, 114, 115, 117], "4": 84, "5": 84, "7": 84, "8": 84, "9": 84, "10": 84, "11": 84, "12": 84, "13": 84, "14": 84, "15": 84, "16": 84, "17": 84, "18": 84, "19": 84, "20": 84, "submenu": 84, "subsubmenu": 84, "structur": 85, "element": 85, "section": 85, "subsect": 85, "subsubsect": 85, "demo": 86, "an": [86, 111], "incred": 86, "via": 87, "git": 87, "creat": [88, 91], "work": [88, 89], "save": [88, 100, 121], "disk": 88, "quickstart": 89, "unsupport": 89, "post": [91, 111], "train": [91, 112, 116], "quantiz": [91, 112, 116], "your": [91, 114, 115, 117], "own": 91, "applic": 91, "directli": 92, "kernel": [93, 96], "plugin": [93, 120], "our": [93, 94, 96], "overload": 94, "metadata": 94, "cross": 95, "import": [95, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113], "within": 96, "test": 96, "wrap": 96, "insert": 96, "cach": [97, 98, 102], "bert": [97, 107, 116], "jit": [98, 118], "aot": [98, 118], "mutabl": 100, "initi": [100, 111], "make": [100, 102], "modif": 100, "stabl": [100, 106], "diffus": [100, 106], "huggingfac": 100, "alloc": 101, "output": [101, 104, 109, 110, 111], "buffer": 101, "measur": 101, "load": [101, 111, 112, 121], "enabl": 101, "disabl": 101, "refit": 102, "new": 102, "standard": 102, "workflow": 102, "refitt": 102, "pretrain": [102, 111], "map": 102, "place": 102, "default": [103, 108], "cleanup": [103, 107], "driver": [103, 107], "error": [103, 107], "note": [103, 107], "gpt2": [104, 109], "necessari": 104, "decod": [104, 109, 110], "sentenc": [104, 109, 110], "resnet": 105, "argument": [105, 107], "avoid": 105, "specifi": 105, "befor": 105, "trt": 105, "cudagraph": [108, 120], "integr": 108, "contain": 108, "break": 108, "llama2": 110, "sam2": 111, "follow": 111, "preprocess": 111, "compon": 111, "process": 111, "visual": 111, "dataset": 112, "loss": 112, "calibr": 112, "tune": 112, "fp8": 112, "stream": 113, "run": 113, "budget": 113, "size": 113, "manag": 113, "serv": [114, 115, 116, 117], "triton": [114, 115, 117], "up": [114, 115, 117], "server": [114, 115, 117], "client": [114, 115, 117], "queri": [114, 115, 117], "notebook": 116, "citrinet": 116, "efficientnet": 116, "mask": 116, "languag": 116, "mlm": 116, "hug": 116, "face": 116, "transform": 116, "acceler": 116, "resnet50": 116, "lenet": 116, "deep": 116, "learn": 116, "object": 116, "detect": 116, "ssd": 116, "int8": 116, "constraint": 118, "mix": 119, "precis": 119, "libtorchtrt": 120, "so": 120, "multi": 120, "safe": 120, "mode": 120, "exportedprogram": 121, "b": 121, "explain": 122, "just": 122, "accept": 122, "return": 122, "ahead": 122, "dla": 123}, "envversion": {"sphinx.domains.c": 2, "sphinx.domains.changeset": 1, "sphinx.domains.citation": 1, "sphinx.domains.cpp": 6, "sphinx.domains.index": 1, "sphinx.domains.javascript": 2, "sphinx.domains.math": 2, "sphinx.domains.python": 3, "sphinx.domains.rst": 2, "sphinx.domains.std": 2, "nbsphinx": 4, "sphinx.ext.intersphinx": 1, "sphinx.ext.todo": 2, "sphinx.ext.viewcode": 1, "sphinx": 56}}) \ No newline at end of file diff --git a/docs/sg_execution_times.html b/docs/sg_execution_times.html index 4c63acfe7f..e2b8a59d6b 100644 --- a/docs/sg_execution_times.html +++ b/docs/sg_execution_times.html @@ -10,7 +10,7 @@ - Computation times — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Computation times — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -746,7 +764,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -762,6 +780,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -800,6 +821,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -813,6 +837,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/src/pytorch-sphinx-theme/docs/changelog.html b/docs/src/pytorch-sphinx-theme/docs/changelog.html index e7d7800497..03fb7a6fc9 100644 --- a/docs/src/pytorch-sphinx-theme/docs/changelog.html +++ b/docs/src/pytorch-sphinx-theme/docs/changelog.html @@ -10,7 +10,7 @@ - Changelog — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Changelog — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -702,7 +720,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -718,6 +736,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -756,6 +777,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -769,6 +793,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/src/pytorch-sphinx-theme/docs/configuring.html b/docs/src/pytorch-sphinx-theme/docs/configuring.html index cc64f58bbf..a980d4ce63 100644 --- a/docs/src/pytorch-sphinx-theme/docs/configuring.html +++ b/docs/src/pytorch-sphinx-theme/docs/configuring.html @@ -10,7 +10,7 @@ - Configuration — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Configuration — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -805,7 +823,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -821,6 +839,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -859,6 +880,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -872,6 +896,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/src/pytorch-sphinx-theme/docs/demo/api.html b/docs/src/pytorch-sphinx-theme/docs/demo/api.html index 19fa661195..c87155fa45 100644 --- a/docs/src/pytorch-sphinx-theme/docs/demo/api.html +++ b/docs/src/pytorch-sphinx-theme/docs/demo/api.html @@ -10,7 +10,7 @@ - 5. :mod:`test_py_module` — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + 5. :mod:`test_py_module` — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -756,7 +774,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -772,6 +790,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -810,6 +831,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -823,6 +847,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/src/pytorch-sphinx-theme/docs/demo/demo.html b/docs/src/pytorch-sphinx-theme/docs/demo/demo.html index 37e93c599d..4bc0b53cad 100644 --- a/docs/src/pytorch-sphinx-theme/docs/demo/demo.html +++ b/docs/src/pytorch-sphinx-theme/docs/demo/demo.html @@ -12,7 +12,7 @@ - 3. Paragraph Level Markup — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + 3. Paragraph Level Markup — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -668,7 +686,7 @@

    3.4.4.

    3.4.5. Code Blocks

    # parsed-literal test
    -curl -O http://someurl/release-v2.6.0.dev0+70e2a38.tar-gz
    +curl -O http://someurl/release-v2.7.0.dev0+d6be4ba.tar-gz
    Code Blocks can have captions.
    {
    @@ -696,7 +714,7 @@ 

    3.4.5.

    3.4.5.1. Emphasized lines with line numbers

    -
    1def some_function():
    +
    1def some_function():
     2    interesting = False
     3    print 'This line is highlighted.'
     4    print 'This one is not...'
    @@ -736,7 +754,7 @@ 

    3.5.1. 2"""Test Module for sphinx_rtd_theme.""" 3 4 - 5class Foo: + 5class Foo: 6 7 """Docstring for class Foo. 8 @@ -1312,7 +1330,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1328,6 +1346,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1366,6 +1387,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1379,6 +1403,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/src/pytorch-sphinx-theme/docs/demo/lists_tables.html b/docs/src/pytorch-sphinx-theme/docs/demo/lists_tables.html index 5c362cf7ca..4335a005d4 100644 --- a/docs/src/pytorch-sphinx-theme/docs/demo/lists_tables.html +++ b/docs/src/pytorch-sphinx-theme/docs/demo/lists_tables.html @@ -10,7 +10,7 @@ - 4. Lists & Tables — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + 4. Lists & Tables — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +

    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -698,7 +716,7 @@

    4.1.5.1. 2"""Test Module for sphinx_rtd_theme.""" 3 4 - 5class Foo: + 5class Foo: 6 7 """Docstring for class Foo. 8 @@ -1287,7 +1305,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1303,6 +1321,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1341,6 +1362,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1354,6 +1378,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/src/pytorch-sphinx-theme/docs/demo/long.html b/docs/src/pytorch-sphinx-theme/docs/demo/long.html index 7426d80cc4..e53d9c5f4c 100644 --- a/docs/src/pytorch-sphinx-theme/docs/demo/long.html +++ b/docs/src/pytorch-sphinx-theme/docs/demo/long.html @@ -10,7 +10,7 @@ - 1. Long Sticky Nav — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + 1. Long Sticky Nav — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -966,7 +984,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -982,6 +1000,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1020,6 +1041,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1033,6 +1057,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/src/pytorch-sphinx-theme/docs/demo/structure.html b/docs/src/pytorch-sphinx-theme/docs/demo/structure.html index a8919d29de..a5b54a1786 100644 --- a/docs/src/pytorch-sphinx-theme/docs/demo/structure.html +++ b/docs/src/pytorch-sphinx-theme/docs/demo/structure.html @@ -10,7 +10,7 @@ - 1. Structural Elements — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + 1. Structural Elements — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -835,7 +853,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -851,6 +869,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -889,6 +910,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -902,6 +926,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/src/pytorch-sphinx-theme/docs/index.html b/docs/src/pytorch-sphinx-theme/docs/index.html index 901df6efd8..71e46900b7 100644 --- a/docs/src/pytorch-sphinx-theme/docs/index.html +++ b/docs/src/pytorch-sphinx-theme/docs/index.html @@ -10,7 +10,7 @@ - <no title> — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + <no title> — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -792,7 +810,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -808,6 +826,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -846,6 +867,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -859,6 +883,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/src/pytorch-sphinx-theme/docs/installing.html b/docs/src/pytorch-sphinx-theme/docs/installing.html index 446b41bb46..93315af032 100644 --- a/docs/src/pytorch-sphinx-theme/docs/installing.html +++ b/docs/src/pytorch-sphinx-theme/docs/installing.html @@ -10,7 +10,7 @@ - Installation — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Installation — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -714,7 +732,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -730,6 +748,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -768,6 +789,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -781,6 +805,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/ts/creating_torchscript_module_in_python.html b/docs/ts/creating_torchscript_module_in_python.html index a009e3603a..306def22e0 100644 --- a/docs/ts/creating_torchscript_module_in_python.html +++ b/docs/ts/creating_torchscript_module_in_python.html @@ -10,7 +10,7 @@ - Creating a TorchScript Module — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Creating a TorchScript Module — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -479,30 +497,30 @@

    PyTorch programs are based around Module s which can be used to compose higher level modules. Modules contain a constructor to set up the modules, parameters and sub-modules and a forward function which describes how to use the parameters and submodules when the module is invoked.

    For example, we can define a LeNet module like this:

    -
     1import torch.nn as nn
    - 2import torch.nn.functional as F
    +
     1import torch.nn as nn
    + 2import torch.nn.functional as F
      3
      4
    - 5class LeNetFeatExtractor(nn.Module):
    - 6    def __init__(self):
    + 5class LeNetFeatExtractor(nn.Module):
    + 6    def __init__(self):
      7        super(LeNetFeatExtractor, self).__init__()
      8        self.conv1 = nn.Conv2d(1, 6, 3)
      9        self.conv2 = nn.Conv2d(6, 16, 3)
     10
    -11    def forward(self, x):
    +11    def forward(self, x):
     12        x = F.max_pool2d(F.relu(self.conv1(x)), (2, 2))
     13        x = F.max_pool2d(F.relu(self.conv2(x)), 2)
     14        return x
     15
     16
    -17class LeNetClassifier(nn.Module):
    -18    def __init__(self):
    +17class LeNetClassifier(nn.Module):
    +18    def __init__(self):
     19        super(LeNetClassifier, self).__init__()
     20        self.fc1 = nn.Linear(16 * 6 * 6, 120)
     21        self.fc2 = nn.Linear(120, 84)
     22        self.fc3 = nn.Linear(84, 10)
     23
    -24    def forward(self, x):
    +24    def forward(self, x):
     25        x = torch.flatten(x, 1)
     26        x = F.relu(self.fc1(x))
     27        x = F.relu(self.fc2(x))
    @@ -510,13 +528,13 @@
     29        return x
     30
     31
    -32class LeNet(nn.Module):
    -33    def __init__(self):
    +32class LeNet(nn.Module):
    +33    def __init__(self):
     34        super(LeNet, self).__init__()
     35        self.feat = LeNetFeatExtractor()
     36        self.classifier = LeNetClassifier()
     37
    -38    def forward(self, x):
    +38    def forward(self, x):
     39        x = self.feat(x)
     40        x = self.classifier(x)
     41        return x
    @@ -529,7 +547,7 @@
     

    From here are two pathways for going from PyTorch Python code to TorchScript code: Tracing and Scripting.

    Tracing follows the path of execution when the module is called and records what happens. To trace an instance of our LeNet module, we can call torch.jit.trace with an example input.

    -
    import torch
    +
    import torch
     
     model = LeNet()
     input_data = torch.empty([1, 1, 32, 32])
    @@ -539,7 +557,7 @@
     

    Scripting actually inspects your code with a compiler and generates an equivalent TorchScript program. The difference is that since tracing is following the execution of your module, it cannot pick up control flow for instance. By working from the Python code, the compiler can include these components. We can run the script compiler on our LeNet module by calling torch.jit.script

    -
    import torch
    +
    import torch
     
     model = LeNet()
     script_model = torch.jit.script(model)
    @@ -581,7 +599,7 @@
     

    Saving TorchScript Module to Disk

    For either traced or scripted modules, you can save the module to disk with the following command

    -
    import torch
    +
    import torch
     
     model = LeNet()
     script_model = torch.jit.script(model)
    @@ -829,7 +847,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -845,6 +863,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -883,6 +904,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -896,6 +920,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/ts/getting_started_with_cpp_api.html b/docs/ts/getting_started_with_cpp_api.html index 046f0487ab..4b7a56729c 100644 --- a/docs/ts/getting_started_with_cpp_api.html +++ b/docs/ts/getting_started_with_cpp_api.html @@ -10,7 +10,7 @@ - Using Torch-TensorRT in C++ — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Using Torch-TensorRT in C++ — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -1007,7 +1025,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1023,6 +1041,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1061,6 +1082,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1074,6 +1098,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/ts/getting_started_with_python_api.html b/docs/ts/getting_started_with_python_api.html index 3dfe059d66..6f23af762a 100644 --- a/docs/ts/getting_started_with_python_api.html +++ b/docs/ts/getting_started_with_python_api.html @@ -10,7 +10,7 @@ - Using Torch-TensorRT in Python — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Using Torch-TensorRT in Python — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -483,7 +501,7 @@ to Torch-TensorRT and you will be returned an optimized TorchScript module to run or add into another PyTorch module. Inputs is a list of torch_tensorrt.Input classes which define input Tensors’ shape, datatype and memory format. Alternatively, if your input is a more complex data type, such as a tuple or list of Tensors, you can use the input_signature argument to specify a collection-based input, such as (List[Tensor], Tuple[Tensor, Tensor]). See the second sample below for an example. You can also specify settings such as operating precision for the engine or target device. After compilation you can save the module just like any other module to load in a deployment application. In order to load a TensorRT/TorchScript module, make sure you first import torch_tensorrt.

    -
    import torch_tensorrt
    +
    import torch_tensorrt
     
     ...
     
    @@ -509,7 +527,7 @@
     
    # Sample using collection-based inputs via the input_signature argument
    -import torch_tensorrt
    +import torch_tensorrt
     
     ...
     
    @@ -534,8 +552,8 @@
     
    # Deployment application
    -import torch
    -import torch_tensorrt
    +import torch
    +import torch_tensorrt
     
     trt_ts_module = torch.jit.load("trt_ts_module.ts")
     input_data = input_data.to("cuda").half()
    @@ -782,7 +800,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -798,6 +816,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -836,6 +857,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -849,6 +873,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/ts/ptq.html b/docs/ts/ptq.html index 023bd14ede..a6b8abac15 100644 --- a/docs/ts/ptq.html +++ b/docs/ts/ptq.html @@ -10,7 +10,7 @@ - Post Training Quantization (PTQ) — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Post Training Quantization (PTQ) — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -890,7 +908,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -906,6 +924,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -944,6 +965,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -957,6 +981,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/ts/torchscript_frontend_from_pytorch.html b/docs/ts/torchscript_frontend_from_pytorch.html index 2d27af3e57..f2584d9c92 100644 --- a/docs/ts/torchscript_frontend_from_pytorch.html +++ b/docs/ts/torchscript_frontend_from_pytorch.html @@ -10,7 +10,7 @@ - Using Torch-TensorRT TorchScript Frontend Directly From PyTorch — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Using Torch-TensorRT TorchScript Frontend Directly From PyTorch — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -474,12 +492,12 @@

    You will now be able to directly access TensorRT from PyTorch APIs. The process to use this feature is very similar to the compilation workflow described in Using Torch-TensorRT in Python

    Start by loading torch_tensorrt into your application.

    -
    import torch
    -import torch_tensorrt
    +
    import torch
    +import torch_tensorrt
     

    Then given a TorchScript module, you can compile it with TensorRT using the torch._C._jit_to_backend("tensorrt", ...) API.

    -
    import torchvision.models as models
    +
    import torchvision.models as models
     
     model = models.mobilenet_v2(pretrained=True)
     script_model = torch.jit.script(model)
    @@ -749,7 +767,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -765,6 +783,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -803,6 +824,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -816,6 +840,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/tutorials/_rendered_examples/dynamo/auto_generate_converters.html b/docs/tutorials/_rendered_examples/dynamo/auto_generate_converters.html index 0f828c1ecc..8b9d00e349 100644 --- a/docs/tutorials/_rendered_examples/dynamo/auto_generate_converters.html +++ b/docs/tutorials/_rendered_examples/dynamo/auto_generate_converters.html @@ -10,7 +10,7 @@ - Automatically Generate a Converter for a Custom Kernel — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Automatically Generate a Converter for a Custom Kernel — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -503,17 +521,17 @@

    Writing Custom Operators in PyTorch
    from typing import Tuple
    +
    @@ -568,7 +586,7 @@

    Writing Plugins for TensorRT using the Quick Deploy Plugin system

    @trtp.impl("torchtrt_ex::elementwise_mul")
    -def _(
    +def _(
         x: trtp.Tensor,
         y: trtp.Tensor,
         b: float,
    @@ -614,11 +632,11 @@ 

    Using our converter with a model
    class MyModel(torch.nn.Module):  # type: ignore[misc]
    -    def __init__(self):
    +
    diff --git a/docs/tutorials/_rendered_examples/dynamo/converter_overloading.html b/docs/tutorials/_rendered_examples/dynamo/converter_overloading.html index 92df70e149..865dbffe17 100644 --- a/docs/tutorials/_rendered_examples/dynamo/converter_overloading.html +++ b/docs/tutorials/_rendered_examples/dynamo/converter_overloading.html @@ -10,7 +10,7 @@ - Overloading Torch-TensorRT Converters with Custom Converters — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Overloading Torch-TensorRT Converters with Custom Converters — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +

    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +

    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us

    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -483,21 +501,21 @@ Torch-TensorRT would normally use.

    In this tutorial, we will demonstrate how to overload Torch-TensorRT’s conversion of the torch.nn.functional.gelu operation to TensorRT with a custom converter that uses a different implementation of the GeLU layer.

    -
    import logging
    -import sys
    +
    import logging
    +import sys
     
    -import torch
    -import torch_tensorrt
    +import torch
    +import torch_tensorrt
     

    GeLU has 2 modes in PyTorch, one using the erf function and the other using the tanh approximation. TensorRT natively supports both implementations as an activation layer, but suppose we want to use a custom implementation of GeLU in TensorRT only for tanh mode.

    -
    class GeLU(torch.nn.Module):
    -    def __init__(self, mode="tanh"):
    +
    class GeLU(torch.nn.Module):
    +    def __init__(self, mode="tanh"):
             super().__init__()
             self.mode = mode
     
    -    def forward(self, x):
    +    def forward(self, x):
             return torch.nn.functional.gelu(x, approximate=self.mode)
     
     
    @@ -519,13 +537,13 @@ 

    Writing a Custom Converter@torch_tensorrt.dynamo.conversion.dynamo_tensorrt_converter decorator. At a code level, converter takes the current conversion state (ConversionCtx), the next operator in the graph to convert, and the arguments to that node and returns the placeholder outputs for that operation, while as side-effect inserting the necessary TensorRT layers into the TensorRT network.

    -
    from typing import Dict, Sequence, Tuple, Union
    +
    from typing import Dict, Sequence, Tuple, Union
     
    -from torch.fx.node import Argument, Node, Target
    -from torch_tensorrt.dynamo import CompilationSettings
    -from torch_tensorrt.dynamo.conversion import ConversionContext
    +from torch.fx.node import Argument, Node, Target
    +from torch_tensorrt.dynamo import CompilationSettings
    +from torch_tensorrt.dynamo.conversion import ConversionContext
     
    -import tensorrt as trt
    +import tensorrt as trt
     
    @@ -567,7 +585,7 @@

    Converter Implementationtorch_tensorrt.dynamo.conversion.impl module and are designed to be composable and interoperable with raw-TensorRT implementations. In this case, we will use the Torch-TensorRT mul, add and tanh functions from impl to implement our alternative GeLU layer.

    -
    def aten_ops_gelu(
    +
    diff --git a/docs/tutorials/_rendered_examples/dynamo/cross_runtime_compilation_for_windows.html b/docs/tutorials/_rendered_examples/dynamo/cross_runtime_compilation_for_windows.html index 55efd182d2..b454fa84ea 100644 --- a/docs/tutorials/_rendered_examples/dynamo/cross_runtime_compilation_for_windows.html +++ b/docs/tutorials/_rendered_examples/dynamo/cross_runtime_compilation_for_windows.html @@ -10,7 +10,7 @@ - Cross runtime compilation for windows example — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Cross runtime compilation for windows example — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +

    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us

    @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -487,12 +505,12 @@

    Cross runtime compilation for windows example

    Imports and Model Definition

    -
    import argparse
    -import platform
    +
    import argparse
    +import platform
     
    -import torch
    -import torch_tensorrt as torchtrt
    -import torchvision.models as models
    +import torch
    +import torch_tensorrt as torchtrt
    +import torchvision.models as models
     
     PARSER = argparse.ArgumentParser(
         description="Cross runtime comilation for windows example: Resnet Model"
    @@ -789,7 +807,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -805,6 +823,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -843,6 +864,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -856,6 +880,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/tutorials/_rendered_examples/dynamo/custom_kernel_plugins.html b/docs/tutorials/_rendered_examples/dynamo/custom_kernel_plugins.html index a71dc1e448..6657ad6a3b 100644 --- a/docs/tutorials/_rendered_examples/dynamo/custom_kernel_plugins.html +++ b/docs/tutorials/_rendered_examples/dynamo/custom_kernel_plugins.html @@ -10,7 +10,7 @@ - Using Custom Kernels within TensorRT Engines with Torch-TensorRT — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Using Custom Kernels within TensorRT Engines with Torch-TensorRT — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +

    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -497,18 +515,18 @@

    Writing Custom Operators in PyTorchOpenAI Triton

    When using custom kernels with PyTorch, it is recommended to take the additional step of registering them as formal operators in PyTorch. This will both make it easier to handle the operation in Torch-TensorRT and simplify its use in PyTorch. This could either be done as part of a C++ library or in Python. (see: Custom ops in C++ and Python custom ops for more details )

    -
    from typing import Any, Sequence
    +
    from typing import Any, Sequence
     
    -import numpy as np
    -import torch
    -import triton
    -import triton.language as tl
    -from torch.library import custom_op
    +import numpy as np
    +import torch
    +import triton
    +import triton.language as tl
    +from torch.library import custom_op
     
     
     # Defining the kernel to be run on the GPU
     @triton.jit  # type: ignore
    -def circ_pad_kernel(
    +def circ_pad_kernel(
         X: torch.Tensor,
         all_pads_0: tl.int32,
         all_pads_2: tl.int32,
    @@ -556,7 +574,7 @@ 

    Writing Custom Operators in PyTorch# The launch code wrapped to expose it as a custom operator in our namespace @custom_op("torchtrt_ex::triton_circular_pad", mutates_args=()) # type: ignore[misc] -def triton_circular_pad(x: torch.Tensor, padding: Sequence[int]) -> torch.Tensor: +def triton_circular_pad(x: torch.Tensor, padding: Sequence[int]) -> torch.Tensor: out_dims = np.ones(len(x.shape), dtype=np.int32) for i in range(np.size(padding) // 2): out_dims[len(out_dims) - i - 1] = ( @@ -633,7 +651,7 @@

    Testing our custom ophere). In our case we can just use the native circular pad operation as our FakeTensor implementation.

    @torch.library.register_fake("torchtrt_ex::triton_circular_pad")  # type: ignore[misc]
    -def _(x: torch.Tensor, padding: Sequence[int]) -> torch.Tensor:
    +def _(x: torch.Tensor, padding: Sequence[int]) -> torch.Tensor:
         return torch.nn.functional.pad(x, padding, "circular")
     
     
    @@ -644,19 +662,19 @@ 

    Testing our custom op

    Using the Custom Operator in a Model

    We can now create models using our custom op. Here is a small example one that uses both natively supported operators (Convolution) and our custom op.

    -
    from typing import Sequence
    +
    from typing import Sequence
     
    -from torch import nn
    +from torch import nn
     
     
    -class MyModel(nn.Module):  # type: ignore[misc]
    -    def __init__(self, padding: Sequence[int]):
    +class MyModel(nn.Module):  # type: ignore[misc]
    +    def __init__(self, padding: Sequence[int]):
             super().__init__()
     
             self.padding = padding
             self.conv = nn.Conv2d(1, 5, kernel_size=3)
     
    -    def forward(self, x: torch.Tensor) -> torch.Tensor:
    +    def forward(self, x: torch.Tensor) -> torch.Tensor:
             padded_x = torch.ops.torchtrt_ex.triton_circular_pad(x, self.padding)
             y = self.conv(padded_x)
     
    @@ -689,7 +707,7 @@ 

    Using the Custom Operator in a Model
    import torch_tensorrt as torchtrt
    +
    import torch_tensorrt as torchtrt
     
     torchtrt.compile(
         my_model,
    @@ -748,17 +766,17 @@ 

    Wrapping Custom Kernels to use in TensorRThere. From a high level, similar to PyTorch you will need to define systems to handle setting up the operator, calculating the effect of the operation abstractly, serializing the op and the actual mechanics of calling the implementation of the op in the engine.

    -
    import pickle as pkl
    -from typing import Any, List, Optional, Self
    +
    import pickle as pkl
    +from typing import Any, List, Optional, Self
     
    -import cupy as cp  # Needed to work around API gaps in PyTorch to build torch.Tensors around preallocated CUDA memory
    -import numpy as np
    +import cupy as cp  # Needed to work around API gaps in PyTorch to build torch.Tensors around preallocated CUDA memory
    +import numpy as np
     
    -import tensorrt as trt
    +import tensorrt as trt
     
     
    -class CircularPaddingPlugin(trt.IPluginV2DynamicExt):  # type: ignore[misc]
    -    def __init__(
    +class CircularPaddingPlugin(trt.IPluginV2DynamicExt):  # type: ignore[misc]
    +    def __init__(
             self, field_collection: Optional[List[trt.PluginFieldCollection]] = None
         ):
             super().__init__()
    @@ -774,12 +792,12 @@ 

    Wrapping Custom Kernels to use in TensorRTassert field_collection[0].name == "pads" self.pads = field_collection[0].data - def get_output_datatype( + def get_output_datatype( self, index: int, input_types: List[trt.DataType] ) -> trt.DataType: return input_types[0] - def get_output_dimensions( + def get_output_dimensions( self, output_index: int, inputs: List[trt.DimsExprs], @@ -797,7 +815,7 @@

    Wrapping Custom Kernels to use in TensorRTreturn output_dims - def configure_plugin( + def configure_plugin( self, inp: List[trt.DynamicPluginTensorDesc], out: List[trt.DynamicPluginTensorDesc], @@ -807,10 +825,10 @@

    Wrapping Custom Kernels to use in TensorRTfor i in range(len(X_dims)): self.X_shape[i] = X_dims[i] - def serialize(self) -> bytes: + def serialize(self) -> bytes: return pkl.dumps({"pads": self.pads}) - def supports_format_combination( + def supports_format_combination( self, pos: int, in_out: List[trt.PluginTensorDesc], num_inputs: int ) -> bool: assert num_inputs == 1 @@ -832,7 +850,7 @@

    Wrapping Custom Kernels to use in TensorRTreturn False - def enqueue( + def enqueue( self, input_desc: List[trt.PluginTensorDesc], output_desc: List[trt.PluginTensorDesc], @@ -900,14 +918,14 @@

    Wrapping Custom Kernels to use in TensorRTBLOCK_SIZE=256, ) - def clone(self) -> Self: + def clone(self) -> Self: cloned_plugin = CircularPaddingPlugin() cloned_plugin.__dict__.update(self.__dict__) return cloned_plugin -class CircularPaddingPluginCreator(trt.IPluginCreator): # type: ignore[misc] - def __init__(self): +class CircularPaddingPluginCreator(trt.IPluginCreator): # type: ignore[misc] + def __init__(self): super().__init__() self.name = "CircularPaddingPlugin" @@ -917,12 +935,12 @@

    Wrapping Custom Kernels to use in TensorRT[trt.PluginField("pads", np.array([]), trt.PluginFieldType.INT32)] ) - def create_plugin( + def create_plugin( self, name: str, field_collection: trt.PluginFieldCollection_ ) -> CircularPaddingPlugin: return CircularPaddingPlugin(field_collection) - def deserialize_plugin(self, name: str, data: bytes) -> CircularPaddingPlugin: + def deserialize_plugin(self, name: str, data: bytes) -> CircularPaddingPlugin: pads_dict = pkl.loads(data) print(pads_dict) deserialized = CircularPaddingPlugin() @@ -941,15 +959,15 @@

    Wrapping Custom Kernels to use in TensorRT

    Now with our TensorRT plugin, we can create a converter so that Torch-TensorRT knows to insert our plugin in place of our custom circular padding operator. More information on writing converters can be found here

    -
    from typing import Dict, Tuple
    +
    from typing import Dict, Tuple
     
    -from torch.fx.node import Argument, Target
    -from torch_tensorrt.dynamo.conversion import (
    +from torch.fx.node import Argument, Target
    +from torch_tensorrt.dynamo.conversion import (
         ConversionContext,
         dynamo_tensorrt_converter,
     )
    -from torch_tensorrt.dynamo.conversion.converter_utils import get_trt_tensor
    -from torch_tensorrt.fx.converters.converter_utils import set_layer_name
    +from torch_tensorrt.dynamo.conversion.converter_utils import get_trt_tensor
    +from torch_tensorrt.fx.converters.converter_utils import set_layer_name
     
     
     @dynamo_tensorrt_converter(
    @@ -957,7 +975,7 @@ 

    Using Torch-TensorRT to Insert the Kernel) # type: ignore # Recall the schema defined above: # torch.ops.torchtrt_ex.triton_circular_pad.default(Tensor x, IntList padding) -> Tensor -def circular_padding_converter( +def circular_padding_converter( ctx: ConversionContext, target: Target, args: Tuple[Argument, ...], @@ -1349,7 +1367,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1365,6 +1383,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1403,6 +1424,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1416,6 +1440,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/tutorials/_rendered_examples/dynamo/engine_caching_bert_example.html b/docs/tutorials/_rendered_examples/dynamo/engine_caching_bert_example.html index bba8633a0e..91d2c3a261 100644 --- a/docs/tutorials/_rendered_examples/dynamo/engine_caching_bert_example.html +++ b/docs/tutorials/_rendered_examples/dynamo/engine_caching_bert_example.html @@ -10,7 +10,7 @@ - Engine Caching (BERT) — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Engine Caching (BERT) — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +

    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -479,11 +497,11 @@

    Engine Caching (BERT)

    Small caching example on BERT.

    -
    import numpy as np
    -import torch
    -import torch_tensorrt
    -from engine_caching_example import remove_timing_cache
    -from transformers import BertModel
    +
    import numpy as np
    +import torch
    +import torch_tensorrt
    +from engine_caching_example import remove_timing_cache
    +from transformers import BertModel
     
     np.random.seed(0)
     torch.manual_seed(0)
    @@ -495,7 +513,7 @@
     ]
     
     
    -def compile_bert(iterations=3):
    +def compile_bert(iterations=3):
         times = []
         start = torch.cuda.Event(enable_timing=True)
         end = torch.cuda.Event(enable_timing=True)
    @@ -795,7 +813,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -811,6 +829,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -849,6 +870,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -862,6 +886,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/tutorials/_rendered_examples/dynamo/engine_caching_example.html b/docs/tutorials/_rendered_examples/dynamo/engine_caching_example.html index 5e61d420fb..3d28fcdaf8 100644 --- a/docs/tutorials/_rendered_examples/dynamo/engine_caching_example.html +++ b/docs/tutorials/_rendered_examples/dynamo/engine_caching_example.html @@ -10,7 +10,7 @@ - Engine Caching — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Engine Caching — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us

    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -496,15 +514,15 @@

    The example uses a pre-trained ResNet18 model and shows the differences between compilation without caching, with caching enabled, and when reusing cached engines.

    -
    import os
    -from typing import Dict, Optional
    +
    import os
    +from typing import Dict, Optional
     
    -import numpy as np
    -import torch
    -import torch_tensorrt as torch_trt
    -import torchvision.models as models
    -from torch_tensorrt.dynamo._defaults import TIMING_CACHE_PATH
    -from torch_tensorrt.dynamo._engine_cache import BaseEngineCache
    +import numpy as np
    +import torch
    +import torch_tensorrt as torch_trt
    +import torchvision.models as models
    +from torch_tensorrt.dynamo._defaults import TIMING_CACHE_PATH
    +from torch_tensorrt.dynamo._engine_cache import BaseEngineCache
     
     np.random.seed(0)
     torch.manual_seed(0)
    @@ -516,7 +534,7 @@
     use_python_runtime = False
     
     
    -def remove_timing_cache(path=TIMING_CACHE_PATH):
    +def remove_timing_cache(path=TIMING_CACHE_PATH):
         if os.path.exists(path):
             os.remove(path)
     
    @@ -534,7 +552,7 @@

    Engine Caching for JIT Compilationcache_built_engines=True), the engine must be refittable (immutable_weights=False). See Refitting Torch-TensorRT Programs with New Weights for more details.

    -
    def torch_compile(iterations=3):
    +
    def torch_compile(iterations=3):
         times = []
         start = torch.cuda.Event(enable_timing=True)
         end = torch.cuda.Event(enable_timing=True)
    @@ -590,7 +608,7 @@ 

    Engine Caching for AOT Compilation
    def dynamo_compile(iterations=3):
    +
    def dynamo_compile(iterations=3):
         times = []
         start = torch.cuda.Event(enable_timing=True)
         end = torch.cuda.Event(enable_timing=True)
    @@ -660,8 +678,8 @@ 

    Custom Engine CacheRAMEngineCache.

    -
    class RAMEngineCache(BaseEngineCache):
    -    def __init__(
    +
    diff --git a/docs/tutorials/_rendered_examples/dynamo/index.html b/docs/tutorials/_rendered_examples/dynamo/index.html index 1425a3853d..6af08095c5 100644 --- a/docs/tutorials/_rendered_examples/dynamo/index.html +++ b/docs/tutorials/_rendered_examples/dynamo/index.html @@ -10,7 +10,7 @@ - Dependencies — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Dependencies — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +

    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -781,7 +799,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -797,6 +815,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -835,6 +856,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -848,6 +872,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/tutorials/_rendered_examples/dynamo/mutable_torchtrt_module_example.html b/docs/tutorials/_rendered_examples/dynamo/mutable_torchtrt_module_example.html index 6d82ec3285..bece6414a9 100644 --- a/docs/tutorials/_rendered_examples/dynamo/mutable_torchtrt_module_example.html +++ b/docs/tutorials/_rendered_examples/dynamo/mutable_torchtrt_module_example.html @@ -10,7 +10,7 @@ - Mutable Torch TensorRT Module — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Mutable Torch TensorRT Module — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +

    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -486,10 +504,10 @@ 1. Sample workflow of Mutable Torch TensorRT Module with ResNet 18 2. Save a Mutable Torch TensorRT Module 3. Integration with Huggingface pipeline in LoRA use case

    -
    import numpy as np
    -import torch
    -import torch_tensorrt as torch_trt
    -import torchvision.models as models
    +
    import numpy as np
    +import torch
    +import torch_tensorrt as torch_trt
    +import torchvision.models as models
     
     np.random.seed(5)
     torch.manual_seed(5)
    @@ -542,7 +560,7 @@ 

    Saving Mutable Torch TensorRT Module

    # The LoRA checkpoint is from https://civitai.com/models/12597/moxin
     
    -from diffusers import DiffusionPipeline
    +from diffusers import DiffusionPipeline
     
     with torch.no_grad():
         settings = {
    @@ -840,7 +858,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -856,6 +874,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -894,6 +915,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -907,6 +931,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/tutorials/_rendered_examples/dynamo/pre_allocated_output_example.html b/docs/tutorials/_rendered_examples/dynamo/pre_allocated_output_example.html index 9e2260de37..ebcac0814f 100644 --- a/docs/tutorials/_rendered_examples/dynamo/pre_allocated_output_example.html +++ b/docs/tutorials/_rendered_examples/dynamo/pre_allocated_output_example.html @@ -10,7 +10,7 @@ - Pre-allocated output buffer — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Pre-allocated output buffer — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -508,18 +526,18 @@

    Imports and Model Definition

    -
    import timeit
    +
    import timeit
     
    -import numpy as np
    -import torch
    -import torch_tensorrt
    -from transformers import BertModel
    +import numpy as np
    +import torch
    +import torch_tensorrt
    +from transformers import BertModel
     

    Define function to measure inference performance

    -
    def test_module_perf(model, *input):
    +
    def test_module_perf(model, *input):
         timings = []
     
         # Warm-up phase to ensure consistent and accurate performance measurements.
    @@ -851,7 +869,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -867,6 +885,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -905,6 +926,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -918,6 +942,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/tutorials/_rendered_examples/dynamo/refit_engine_example.html b/docs/tutorials/_rendered_examples/dynamo/refit_engine_example.html index 6d1d2057e7..1b9abcffb6 100644 --- a/docs/tutorials/_rendered_examples/dynamo/refit_engine_example.html +++ b/docs/tutorials/_rendered_examples/dynamo/refit_engine_example.html @@ -10,7 +10,7 @@ - Refitting Torch-TensorRT Programs with New Weights — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Refitting Torch-TensorRT Programs with New Weights — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +

    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -500,11 +518,11 @@

    Standard Workflow

    Imports and model definition

    -
    import numpy as np
    -import torch
    -import torch_tensorrt as torch_trt
    -import torchvision.models as models
    -from torch_tensorrt.dynamo import refit_module_weights
    +
    import numpy as np
    +import torch
    +import torch_tensorrt as torch_trt
    +import torchvision.models as models
    +from torch_tensorrt.dynamo import refit_module_weights
     
     np.random.seed(0)
     torch.manual_seed(0)
    @@ -862,7 +880,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -878,6 +896,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -916,6 +937,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -929,6 +953,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/tutorials/_rendered_examples/dynamo/torch_compile_advanced_usage.html b/docs/tutorials/_rendered_examples/dynamo/torch_compile_advanced_usage.html index 8a8d81d53e..56fa21e55e 100644 --- a/docs/tutorials/_rendered_examples/dynamo/torch_compile_advanced_usage.html +++ b/docs/tutorials/_rendered_examples/dynamo/torch_compile_advanced_usage.html @@ -10,7 +10,7 @@ - Torch Compile Advanced Usage — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Torch Compile Advanced Usage — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us

    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -481,17 +499,17 @@

    This interactive script is intended as an overview of the process by which torch_tensorrt.compile(…, ir=”torch_compile”, …) works, and how it integrates with the torch.compile API.

    Imports and Model Definition

    -
    import torch
    -import torch_tensorrt
    +
    import torch
    +import torch_tensorrt
     
    # We begin by defining a model
    -class Model(torch.nn.Module):
    -    def __init__(self) -> None:
    +class Model(torch.nn.Module):
    +    def __init__(self) -> None:
             super().__init__()
             self.relu = torch.nn.ReLU()
     
    -    def forward(self, x: torch.Tensor, y: torch.Tensor):
    +    def forward(self, x: torch.Tensor, y: torch.Tensor):
             x_out = self.relu(x)
             y_out = self.relu(y)
             x_y_out = x_out + y_out
    @@ -829,7 +847,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -845,6 +863,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -883,6 +904,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -896,6 +920,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/tutorials/_rendered_examples/dynamo/torch_compile_gpt2.html b/docs/tutorials/_rendered_examples/dynamo/torch_compile_gpt2.html index eef0527884..6652b1696c 100644 --- a/docs/tutorials/_rendered_examples/dynamo/torch_compile_gpt2.html +++ b/docs/tutorials/_rendered_examples/dynamo/torch_compile_gpt2.html @@ -10,7 +10,7 @@ - Compiling GPT2 using the Torch-TensorRT torch.compile frontend — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Compiling GPT2 using the Torch-TensorRT torch.compile frontend — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +

    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -487,9 +505,9 @@ get the graph module representation of the graph. Torch-TensorRT converts this graph into an optimized TensorRT engine.

    Import necessary libraries

    -
    import torch
    -import torch_tensorrt
    -from transformers import AutoModelForCausalLM, AutoTokenizer
    +
    import torch
    +import torch_tensorrt
    +from transformers import AutoModelForCausalLM, AutoTokenizer
     
    @@ -848,7 +866,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -864,6 +882,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -902,6 +923,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -915,6 +939,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/tutorials/_rendered_examples/dynamo/torch_compile_resnet_example.html b/docs/tutorials/_rendered_examples/dynamo/torch_compile_resnet_example.html index 633b504bb3..dc33d782f9 100644 --- a/docs/tutorials/_rendered_examples/dynamo/torch_compile_resnet_example.html +++ b/docs/tutorials/_rendered_examples/dynamo/torch_compile_resnet_example.html @@ -10,7 +10,7 @@ - Compiling ResNet with dynamic shapes using the torch.compile backend — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Compiling ResNet with dynamic shapes using the torch.compile backend — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -481,9 +499,9 @@

    This interactive script is intended as a sample of the Torch-TensorRT workflow with torch.compile on a ResNet model.

    Imports and Model Definition

    -
    import torch
    -import torch_tensorrt
    -import torchvision.models as models
    +
    import torch
    +import torch_tensorrt
    +import torchvision.models as models
     
    # Initialize model with half precision and sample inputs
    @@ -843,7 +861,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -859,6 +877,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -897,6 +918,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -910,6 +934,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/tutorials/_rendered_examples/dynamo/torch_compile_stable_diffusion.html b/docs/tutorials/_rendered_examples/dynamo/torch_compile_stable_diffusion.html index c30e8fa19a..3ead6247c7 100644 --- a/docs/tutorials/_rendered_examples/dynamo/torch_compile_stable_diffusion.html +++ b/docs/tutorials/_rendered_examples/dynamo/torch_compile_stable_diffusion.html @@ -10,7 +10,7 @@ - Compiling Stable Diffusion model using the torch.compile backend — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Compiling Stable Diffusion model using the torch.compile backend — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -482,9 +500,9 @@ ../../../_images/majestic_castle.png

    Imports and Model Definition

    -
    import torch
    -import torch_tensorrt
    -from diffusers import DiffusionPipeline
    +
    import torch
    +import torch_tensorrt
    +from diffusers import DiffusionPipeline
     
     model_id = "CompVis/stable-diffusion-v1-4"
     device = "cuda:0"
    @@ -773,7 +791,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -789,6 +807,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -827,6 +848,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -840,6 +864,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/tutorials/_rendered_examples/dynamo/torch_compile_transformers_example.html b/docs/tutorials/_rendered_examples/dynamo/torch_compile_transformers_example.html index c7b2b9a0b7..33b2385aae 100644 --- a/docs/tutorials/_rendered_examples/dynamo/torch_compile_transformers_example.html +++ b/docs/tutorials/_rendered_examples/dynamo/torch_compile_transformers_example.html @@ -10,7 +10,7 @@ - Compiling BERT using the torch.compile backend — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Compiling BERT using the torch.compile backend — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -481,9 +499,9 @@

    This interactive script is intended as a sample of the Torch-TensorRT workflow with torch.compile on a BERT model.

    Imports and Model Definition

    -
    import torch
    -import torch_tensorrt
    -from transformers import BertModel
    +
    import torch
    +import torch_tensorrt
    +from transformers import BertModel
     
    # Initialize model with float precision and sample inputs
    @@ -832,7 +850,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -848,6 +866,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -886,6 +907,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -899,6 +923,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/tutorials/_rendered_examples/dynamo/torch_export_cudagraphs.html b/docs/tutorials/_rendered_examples/dynamo/torch_export_cudagraphs.html index 877c7701af..6c63cb1a97 100644 --- a/docs/tutorials/_rendered_examples/dynamo/torch_export_cudagraphs.html +++ b/docs/tutorials/_rendered_examples/dynamo/torch_export_cudagraphs.html @@ -10,7 +10,7 @@ - Torch Export with Cudagraphs — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Torch Export with Cudagraphs — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -481,9 +499,9 @@

    This interactive script is intended as an overview of the process by which the Torch-TensorRT Cudagraphs integration can be used in the ir=”dynamo” path. The functionality works similarly in the torch.compile path as well.

    Imports and Model Definition

    -
    import torch
    -import torch_tensorrt
    -import torchvision.models as models
    +
    import torch
    +import torch_tensorrt
    +import torchvision.models as models
     
    @@ -550,8 +568,8 @@

    Cuda graphs with module that contains graph breaks -
    class SampleModel(torch.nn.Module):
    -    def forward(self, x):
    +
    diff --git a/docs/tutorials/_rendered_examples/dynamo/torch_export_gpt2.html b/docs/tutorials/_rendered_examples/dynamo/torch_export_gpt2.html index 6ce47460a0..8e98176b8b 100644 --- a/docs/tutorials/_rendered_examples/dynamo/torch_export_gpt2.html +++ b/docs/tutorials/_rendered_examples/dynamo/torch_export_gpt2.html @@ -10,7 +10,7 @@ - Compiling GPT2 using the dynamo backend — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Compiling GPT2 using the dynamo backend — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -481,10 +499,10 @@

    This script illustrates Torch-TensorRT workflow with dynamo backend on popular GPT2 model.

    Imports and Model Definition

    -
    import torch
    -import torch_tensorrt
    -from transformers import AutoModelForCausalLM, AutoTokenizer
    -from utils import export_llm, generate
    +
    import torch
    +import torch_tensorrt
    +from transformers import AutoModelForCausalLM, AutoTokenizer
    +from utils import export_llm, generate
     
    # Define the parameters and initialize the model
    @@ -823,7 +841,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -839,6 +857,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -877,6 +898,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -890,6 +914,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/tutorials/_rendered_examples/dynamo/torch_export_llama2.html b/docs/tutorials/_rendered_examples/dynamo/torch_export_llama2.html index d24b7872cc..830b37ed46 100644 --- a/docs/tutorials/_rendered_examples/dynamo/torch_export_llama2.html +++ b/docs/tutorials/_rendered_examples/dynamo/torch_export_llama2.html @@ -10,7 +10,7 @@ - Compiling Llama2 using the dynamo backend — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Compiling Llama2 using the dynamo backend — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -481,10 +499,10 @@

    This script illustrates Torch-TensorRT workflow with dynamo backend on popular Llama2 model.

    Imports and Model Definition

    -
    import torch
    -import torch_tensorrt
    -from transformers import AutoModelForCausalLM, AutoTokenizer
    -from utils import export_llm, generate
    +
    import torch
    +import torch_tensorrt
    +from transformers import AutoModelForCausalLM, AutoTokenizer
    +from utils import export_llm, generate
     

    Define the parameters and initialize the model

    @@ -829,7 +847,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -845,6 +863,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -883,6 +904,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -896,6 +920,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/tutorials/_rendered_examples/dynamo/torch_export_sam2.html b/docs/tutorials/_rendered_examples/dynamo/torch_export_sam2.html index 71b3dad2a2..cfbd7fe9ef 100644 --- a/docs/tutorials/_rendered_examples/dynamo/torch_export_sam2.html +++ b/docs/tutorials/_rendered_examples/dynamo/torch_export_sam2.html @@ -10,7 +10,7 @@ - Compiling SAM2 using the dynamo backend — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Compiling SAM2 using the dynamo backend — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -495,15 +513,15 @@

    Import the following libraries

    -
    import matplotlib
    -import matplotlib.pyplot as plt
    -import numpy as np
    -import pandas as pd
    -import torch
    -import torch_tensorrt
    -from PIL import Image
    -from sam2.sam2_image_predictor import SAM2ImagePredictor
    -from sam_components import SAM2FullModel
    +
    import matplotlib
    +import matplotlib.pyplot as plt
    +import numpy as np
    +import pandas as pd
    +import torch
    +import torch_tensorrt
    +from PIL import Image
    +from sam2.sam2_image_predictor import SAM2ImagePredictor
    +from sam_components import SAM2FullModel
     
     matplotlib.use("Agg")
     
    @@ -521,8 +539,8 @@

    Define the SAM2 modelSAM2FullModel which uses these utilities from SAM2ImagePredictor class. SAM2FullModel performs feature extraction and mask prediction in a single step instead of two step process of SAM2ImagePredictor (set_image and predict functions)

    -
    class SAM2FullModel(torch.nn.Module):
    -    def __init__(self, model):
    +
    class SAM2FullModel(torch.nn.Module):
    +    def __init__(self, model):
             super().__init__()
             self.image_encoder = model.forward_image
             self._prepare_backbone_features = model._prepare_backbone_features
    @@ -535,7 +553,7 @@ 

    Define the SAM2 modelself._bb_feat_sizes = [(256, 256), (128, 128), (64, 64)] - def forward(self, image, point_coords, point_labels): + def forward(self, image, point_coords, point_labels): backbone_out = self.image_encoder(image) _, vision_feats, _, _ = self._prepare_backbone_features(backbone_out) @@ -602,7 +620,7 @@

    Preprocessing componentshttps://github.com/facebookresearch/sam2/blob/main/sam2/utils/transforms.py

    -
    def preprocess_inputs(image, predictor):
    +
    def preprocess_inputs(image, predictor):
         w, h = image.size
         orig_hw = [(h, w)]
         input_image = predictor._transforms(np.array(image))[None, ...].to("cuda:0")
    @@ -631,7 +649,7 @@ 

    Preprocessing components

    The following functions implement postprocessing components which include plotting and visualizing masks and points. We use the SAM2Transforms to post process these masks and sort them via confidence score.

    -
    def postprocess_masks(out, predictor, image):
    +
    def postprocess_masks(out, predictor, image):
         """Postprocess low-resolution masks and convert them for visualization."""
         orig_hw = (image.size[1], image.size[0])  # (height, width)
         masks = predictor._transforms.postprocess_masks(out["low_res_masks"], orig_hw)
    @@ -641,7 +659,7 @@ 

    Post Processing componentsreturn masks[sorted_indices], scores[sorted_indices] -def show_mask(mask, ax, random_color=False, borders=True): +def show_mask(mask, ax, random_color=False, borders=True): if random_color: color = np.concatenate([np.random.random(3), np.array([0.6])], axis=0) else: @@ -650,7 +668,7 @@

    Post Processing componentsmask = mask.astype(np.uint8) mask_image = mask.reshape(h, w, 1) * color.reshape(1, 1, -1) if borders: - import cv2 + import cv2 contours, _ = cv2.findContours(mask, cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_NONE) # Try to smooth contours @@ -663,7 +681,7 @@

    Post Processing componentsax.imshow(mask_image) -def show_points(coords, labels, ax, marker_size=375): +def show_points(coords, labels, ax, marker_size=375): pos_points = coords[labels == 1] neg_points = coords[labels == 0] ax.scatter( @@ -686,7 +704,7 @@

    Post Processing components) -def visualize_masks( +def visualize_masks( image, masks, scores, point_coords, point_labels, title_prefix="", save=True ): """Visualize and save masks overlaid on the original image.""" @@ -1016,7 +1034,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1032,6 +1050,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1070,6 +1091,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1083,6 +1107,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/tutorials/_rendered_examples/dynamo/vgg16_ptq.html b/docs/tutorials/_rendered_examples/dynamo/vgg16_ptq.html index 00842e68bd..773269a045 100644 --- a/docs/tutorials/_rendered_examples/dynamo/vgg16_ptq.html +++ b/docs/tutorials/_rendered_examples/dynamo/vgg16_ptq.html @@ -10,7 +10,7 @@ - Deploy Quantized Models using Torch-TensorRT — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Deploy Quantized Models using Torch-TensorRT — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -481,20 +499,20 @@

    Here we demonstrate how to deploy a model quantized to INT8 or FP8 using the Dynamo frontend of Torch-TensorRT

    Imports and Model Definition

    -
    import argparse
    +
    import argparse
     
    -import modelopt.torch.quantization as mtq
    -import torch
    -import torch.nn as nn
    -import torch.nn.functional as F
    -import torch_tensorrt as torchtrt
    -import torchvision.datasets as datasets
    -import torchvision.transforms as transforms
    -from modelopt.torch.quantization.utils import export_torch_mode
    +import modelopt.torch.quantization as mtq
    +import torch
    +import torch.nn as nn
    +import torch.nn.functional as F
    +import torch_tensorrt as torchtrt
    +import torchvision.datasets as datasets
    +import torchvision.transforms as transforms
    +from modelopt.torch.quantization.utils import export_torch_mode
     
     
    -class VGG(nn.Module):
    -    def __init__(self, layer_spec, num_classes=1000, init_weights=False):
    +class VGG(nn.Module):
    +    def __init__(self, layer_spec, num_classes=1000, init_weights=False):
             super(VGG, self).__init__()
     
             layers = []
    @@ -524,7 +542,7 @@ 

    Imports and Model Definitionif init_weights: self._initialize_weights() - def _initialize_weights(self): + def _initialize_weights(self): for m in self.modules(): if isinstance(m, nn.Conv2d): nn.init.kaiming_normal_(m.weight, mode="fan_out", nonlinearity="relu") @@ -537,7 +555,7 @@

    Imports and Model Definitionnn.init.normal_(m.weight, 0, 0.01) nn.init.constant_(m.bias, 0) - def forward(self, x): + def forward(self, x): x = self.features(x) x = self.avgpool(x) x = torch.flatten(x, 1) @@ -545,7 +563,7 @@

    Imports and Model Definitionreturn x -def vgg16(num_classes=1000, init_weights=False): +def vgg16(num_classes=1000, init_weights=False): vgg16_cfg = [ 64, 64, @@ -600,7 +618,7 @@

    Load the pre-trained model weightsweights = ckpt["model_state_dict"] if torch.cuda.device_count() > 1: - from collections import OrderedDict + from collections import OrderedDict new_state_dict = OrderedDict() for k, v in weights.items(): @@ -646,7 +664,7 @@

    Load training dataset and define loss function for PTQ

    Define Calibration Loop for quantization

    -
    def calibrate_loop(model):
    +
    diff --git a/docs/tutorials/_rendered_examples/dynamo/weight_streaming_example.html b/docs/tutorials/_rendered_examples/dynamo/weight_streaming_example.html index bd71e399eb..5926433ce7 100644 --- a/docs/tutorials/_rendered_examples/dynamo/weight_streaming_example.html +++ b/docs/tutorials/_rendered_examples/dynamo/weight_streaming_example.html @@ -10,7 +10,7 @@ - Weight Streaming — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Weight Streaming — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +

    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -497,17 +515,17 @@

    Imports and Model Definition

    -
    import copy
    -import timeit
    +
    import copy
    +import timeit
     
    -import numpy as np
    -import torch
    -import torch_tensorrt
    -from transformers import AutoModelForCausalLM
    -from utils import export_llm
    +import numpy as np
    +import torch
    +import torch_tensorrt
    +from transformers import AutoModelForCausalLM
    +from utils import export_llm
     
     
    -def time_generate(model, inputs, output_seq_length, iterations=10):
    +def time_generate(model, inputs, output_seq_length, iterations=10):
         """
         Measure the time for generating a sentence over certain number of iterations
         """
    @@ -898,7 +916,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -914,6 +932,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -952,6 +973,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -965,6 +989,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/tutorials/_rendered_examples/index.html b/docs/tutorials/_rendered_examples/index.html index b79b764823..b562b08eed 100644 --- a/docs/tutorials/_rendered_examples/index.html +++ b/docs/tutorials/_rendered_examples/index.html @@ -10,7 +10,7 @@ - Torch-TensorRT Tutorials — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Torch-TensorRT Tutorials — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us

    @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -999,7 +1017,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -1015,6 +1033,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -1053,6 +1074,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -1066,6 +1090,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/tutorials/_rendered_examples/triton/index.html b/docs/tutorials/_rendered_examples/triton/index.html index 8a4211673c..e32a5dbdbd 100644 --- a/docs/tutorials/_rendered_examples/triton/index.html +++ b/docs/tutorials/_rendered_examples/triton/index.html @@ -10,7 +10,7 @@ - Serving a Torch-TensorRT model with Triton — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Serving a Torch-TensorRT model with Triton — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -133,8 +133,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -155,6 +155,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +

    @@ -202,6 +205,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -217,7 +224,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -273,7 +291,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -900,7 +918,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -916,6 +934,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -954,6 +975,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -967,6 +991,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/tutorials/notebooks.html b/docs/tutorials/notebooks.html index bd562e28f3..a3ded7a6a0 100644 --- a/docs/tutorials/notebooks.html +++ b/docs/tutorials/notebooks.html @@ -10,7 +10,7 @@ - Legacy notebooks — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Legacy notebooks — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -863,7 +881,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -879,6 +897,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -917,6 +938,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -930,6 +954,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/tutorials/serving_torch_tensorrt_with_triton.html b/docs/tutorials/serving_torch_tensorrt_with_triton.html index acd474e4ec..ff04b70b88 100644 --- a/docs/tutorials/serving_torch_tensorrt_with_triton.html +++ b/docs/tutorials/serving_torch_tensorrt_with_triton.html @@ -10,7 +10,7 @@ - Serving a Torch-TensorRT model with Triton — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Serving a Torch-TensorRT model with Triton — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -911,7 +929,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -927,6 +945,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -965,6 +986,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -978,6 +1002,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/user_guide/dynamic_shapes.html b/docs/user_guide/dynamic_shapes.html index 089455645d..8029fbb06a 100644 --- a/docs/user_guide/dynamic_shapes.html +++ b/docs/user_guide/dynamic_shapes.html @@ -10,7 +10,7 @@ - Dynamic shapes with Torch-TensorRT — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Dynamic shapes with Torch-TensorRT — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -480,8 +498,8 @@

    Dynamic shapes using torch.export (AOT)ir=dynamo with ir=ts and the behavior is exactly the same.

    -
    import torch
    -import torch_tensorrt
    +
    import torch
    +import torch_tensorrt
     
     model = MyModel().eval().cuda()
     # Compile with static shapes
    @@ -518,14 +536,14 @@ 

    Custom Dynamic Shape Constraintsdocumentation to export the Pytorch module with dynamic shapes. Here’s a simple example that exports a matmul layer with some restrictions on dynamic dimensions.

    -
    import torch
    -import torch_tensorrt
    +
    import torch
    +import torch_tensorrt
     
    -class MatMul(torch.nn.Module):
    -    def __init__(self):
    +class MatMul(torch.nn.Module):
    +    def __init__(self):
             super().__init__()
     
    -    def forward(self, query, key):
    +    def forward(self, query, key):
             attn_weight = torch.matmul(query, key.transpose(-1, -2))
             return attn_weight
     
    @@ -546,8 +564,8 @@ 

    Dynamic shapes using torch.compile (JIT)torch_tensorrt.compile(model, inputs, ir="torch_compile") returns a torch.compile boxed function with the backend configured to TensorRT. In the case of ir=torch_compile, users can provide dynamic shape information for the inputs using torch._dynamo.mark_dynamic API (https://pytorch.org/docs/stable/torch.compiler_dynamic_shapes.html) to avoid recompilation of TensorRT engines.

    -
    import torch
    -import torch_tensorrt
    +
    import torch
    +import torch_tensorrt
     
     model = MyModel().eval().cuda()
     inputs = torch.randn((1, 3, 224, 224), dtype=float32)
    @@ -808,7 +826,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -824,6 +842,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -862,6 +883,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -875,6 +899,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/user_guide/mixed_precision.html b/docs/user_guide/mixed_precision.html index aa7697dcdd..644b0d2ec9 100644 --- a/docs/user_guide/mixed_precision.html +++ b/docs/user_guide/mixed_precision.html @@ -10,7 +10,7 @@ - Compile Mixed Precision models with Torch-TensorRT — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Compile Mixed Precision models with Torch-TensorRT — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +

    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -474,14 +492,14 @@

    Compile Mixed Precision models with Torch-TensorRT

    Consider the following Pytorch model which explicitly casts intermediate layer to run in FP16.

    -
    class MyModule(torch.nn.Module):
    -    def __init__(self):
    +
    class MyModule(torch.nn.Module):
    +    def __init__(self):
             super().__init__()
             self.linear1 = torch.nn.Linear(10,10)
             self.linear2 = torch.nn.Linear(10,30).half()
             self.linear3 = torch.nn.Linear(30,40)
     
    -    def forward(self, x):
    +    def forward(self, x):
             x = self.linear1(x)
             x = x.to(torch.float16)
             x = self.linear2(x)
    @@ -769,7 +787,7 @@ 

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -785,6 +803,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -823,6 +844,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -836,6 +860,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/user_guide/runtime.html b/docs/user_guide/runtime.html index 8a2d5cebbd..72c83c617d 100644 --- a/docs/user_guide/runtime.html +++ b/docs/user_guide/runtime.html @@ -10,7 +10,7 @@ - Deploying Torch-TensorRT Programs — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Deploying Torch-TensorRT Programs — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us

    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -799,7 +817,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -815,6 +833,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -853,6 +874,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -866,6 +890,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/user_guide/saving_models.html b/docs/user_guide/saving_models.html index 4db68ce81d..222ef9711b 100644 --- a/docs/user_guide/saving_models.html +++ b/docs/user_guide/saving_models.html @@ -10,7 +10,7 @@ - Saving models compiled with Torch-TensorRT — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Saving models compiled with Torch-TensorRT — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs + @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -486,13 +504,13 @@

    Dynamo IR

    a) ExportedProgram

    Here’s an example usage

    -
    import torch
    -import torch_tensorrt
    +
    import torch
    +import torch_tensorrt
     
     model = MyModel().eval().cuda()
     inputs = [torch.randn((1, 3, 224, 224)).cuda()]
     # trt_ep is a torch.fx.GraphModule object
    -trt_gm = torch_tensorrt.compile(model, ir="dynamo", inputs)
    +trt_gm = torch_tensorrt.compile(model, ir="dynamo", inputs=inputs)
     torch_tensorrt.save(trt_gm, "trt.ep", inputs=inputs)
     
     # Later, you can load it and run inference
    @@ -503,13 +521,13 @@ 

    a) ExportedProgram

    b) Torchscript

    -
    import torch
    -import torch_tensorrt
    +
    import torch
    +import torch_tensorrt
     
     model = MyModel().eval().cuda()
     inputs = [torch.randn((1, 3, 224, 224)).cuda()]
     # trt_gm is a torch.fx.GraphModule object
    -trt_gm = torch_tensorrt.compile(model, ir="dynamo", inputs)
    +trt_gm = torch_tensorrt.compile(model, ir="dynamo", inputs=inputs)
     torch_tensorrt.save(trt_gm, "trt.ts", output_format="torchscript", inputs=inputs)
     
     # Later, you can load it and run inference
    @@ -523,12 +541,12 @@ 

    b) Torchscript

    In Torch-TensorRT 1.X versions, the primary way to compile and run inference with Torch-TensorRT is using Torchscript IR. For ir=ts, this behavior stays the same in 2.X versions as well.

    -
    import torch
    -import torch_tensorrt
    +
    import torch
    +import torch_tensorrt
     
     model = MyModel().eval().cuda()
     inputs = [torch.randn((1, 3, 224, 224)).cuda()]
    -trt_ts = torch_tensorrt.compile(model, ir="ts", inputs) # Output is a ScriptModule object
    +trt_ts = torch_tensorrt.compile(model, ir="ts", inputs=inputs) # Output is a ScriptModule object
     torch.jit.save(trt_ts, "trt_model.ts")
     
     # Later, you can load it and run inference
    @@ -542,8 +560,8 @@ 

    Loading the models
    import torch
    -import torch_tensorrt
    +
    diff --git a/docs/user_guide/torch_tensorrt_explained.html b/docs/user_guide/torch_tensorrt_explained.html index 6a219f31b3..42ca24daec 100644 --- a/docs/user_guide/torch_tensorrt_explained.html +++ b/docs/user_guide/torch_tensorrt_explained.html @@ -10,7 +10,7 @@ - Torch-TensorRT Explained — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + Torch-TensorRT Explained — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +

    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +
    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -852,7 +870,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -868,6 +886,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -906,6 +927,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -919,6 +943,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docs/user_guide/using_dla.html b/docs/user_guide/using_dla.html index 1cdd1c6bbb..0cd04c9c0d 100644 --- a/docs/user_guide/using_dla.html +++ b/docs/user_guide/using_dla.html @@ -10,7 +10,7 @@ - DLA — Torch-TensorRT v2.6.0.dev0+70e2a38 documentation + DLA — Torch-TensorRT v2.7.0.dev0+d6be4ba documentation @@ -135,8 +135,8 @@ Developer Resources

    Find resources and get questions answered

    - - Contributor Awards - 2023 + + Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

    @@ -157,6 +157,9 @@ ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    + + ExecuTorch Docs +
    @@ -204,6 +207,10 @@ Events

    Find events, webinars, and podcasts

    + + Newsletter +

    Stay up-to-date with the latest updates

    +

    @@ -219,7 +226,18 @@ Governing Board -

    +
    + + Cloud Credit Program + + + Technical Advisory Council + + + Staff + + + Contact Us
    @@ -275,7 +293,7 @@
    - v2.6.0.dev0+70e2a38 + v2.7.0.dev0+d6be4ba
    @@ -746,7 +764,7 @@

    Resources

    Developer Resources
  • - Contributor Awards - 2023 + Contributor Awards - 2024
  • @@ -762,6 +780,9 @@

    Resources

  • ExecuTorch
  • +
  • + ExecuTorch Documentation +
  • @@ -800,6 +821,9 @@

    Resources

  • Events
  • +
  • + Newsletter +
  • @@ -813,6 +837,18 @@

    Resources

  • Governing Board
  • +
  • + Cloud Credit Program +
  • +
  • + Technical Advisory Council +
  • +
  • + Staff +
  • +
  • + Contact Us +
  • diff --git a/docsrc/user_guide/saving_models.rst b/docsrc/user_guide/saving_models.rst index 42cff7b954..dc4b5da222 100644 --- a/docsrc/user_guide/saving_models.rst +++ b/docsrc/user_guide/saving_models.rst @@ -34,7 +34,7 @@ Here's an example usage model = MyModel().eval().cuda() inputs = [torch.randn((1, 3, 224, 224)).cuda()] # trt_ep is a torch.fx.GraphModule object - trt_gm = torch_tensorrt.compile(model, ir="dynamo", inputs) + trt_gm = torch_tensorrt.compile(model, ir="dynamo", inputs=inputs) torch_tensorrt.save(trt_gm, "trt.ep", inputs=inputs) # Later, you can load it and run inference @@ -52,7 +52,7 @@ b) Torchscript model = MyModel().eval().cuda() inputs = [torch.randn((1, 3, 224, 224)).cuda()] # trt_gm is a torch.fx.GraphModule object - trt_gm = torch_tensorrt.compile(model, ir="dynamo", inputs) + trt_gm = torch_tensorrt.compile(model, ir="dynamo", inputs=inputs) torch_tensorrt.save(trt_gm, "trt.ts", output_format="torchscript", inputs=inputs) # Later, you can load it and run inference @@ -73,7 +73,7 @@ For `ir=ts`, this behavior stays the same in 2.X versions as well. model = MyModel().eval().cuda() inputs = [torch.randn((1, 3, 224, 224)).cuda()] - trt_ts = torch_tensorrt.compile(model, ir="ts", inputs) # Output is a ScriptModule object + trt_ts = torch_tensorrt.compile(model, ir="ts", inputs=inputs) # Output is a ScriptModule object torch.jit.save(trt_ts, "trt_model.ts") # Later, you can load it and run inference diff --git a/examples/distributed_inference/README.md b/examples/distributed_inference/README.md index 4c88570b6f..d4cf9508e1 100644 --- a/examples/distributed_inference/README.md +++ b/examples/distributed_inference/README.md @@ -14,3 +14,37 @@ See the examples started with `data_parallel` for more details. Here we use torch.distributed as an example, but compilation with tensor parallelism is agnostic to the implementation framework as long as the module is properly sharded. torchrun --nproc_per_node=2 tensor_parallel_llama2.py + +3. Tensor parallel distributed inference using nccl ops plugin + + apt install libmpich-dev + + apt install libopenmpi-dev + + #For python3.10 + + pip install tensorrt-llm + + For other python versions, you need to load the libnvinfer_plugin_tensorrt_llm.so. Please set that in the environment variable export TRTLLM_PLUGINS_PATH={lib_path}. For example, we have already set the variable in initialize_distributed_env(). You can replace this with your TRTLLM_PLUGINS_PATH and unset it there + + #then pip install the tensorrt and torch version compatible with installed torchTRT + + mpirun -n 2 --allow-run-as-root python tensor_parallel_simple_example.py + + #For other python + +4. Tensor parallel distributed llama3 inference using nccl ops plugin + + apt install libmpich-dev + + apt install libopenmpi-dev + +#For python3.10 + + pip install tensorrt-llm + + For other python versions, you need to load the libnvinfer_plugin_tensorrt_llm.so + + #then pip install the tensorrt and torch version compatible with installed torchTRT + + mpirun -n 2 --allow-run-as-root python tensor_parallel_llama3.py diff --git a/examples/distributed_inference/requirement.txt b/examples/distributed_inference/requirement.txt index 6d8e0aa9f2..bb3b0f28f1 100644 --- a/examples/distributed_inference/requirement.txt +++ b/examples/distributed_inference/requirement.txt @@ -1,3 +1,4 @@ accelerate transformers -diffusers \ No newline at end of file +diffusers +tensorrt-llm \ No newline at end of file diff --git a/examples/distributed_inference/tensor_parallel_initialize_dist.py b/examples/distributed_inference/tensor_parallel_initialize_dist.py new file mode 100644 index 0000000000..21e4cbc282 --- /dev/null +++ b/examples/distributed_inference/tensor_parallel_initialize_dist.py @@ -0,0 +1,67 @@ +import logging +import os +from typing import Any, Callable, Dict, List, Optional, Sequence, Tuple, Union + +import numpy as np +import tensorrt as trt +import torch +import torch.distributed as dist +from torch.distributed._tensor.device_mesh import init_device_mesh + + +def find_repo_root(max_depth=10): + dir_path = os.path.dirname(os.path.realpath(__file__)) + for i in range(max_depth): + files = os.listdir(dir_path) + if "MODULE.bazel" in files: + return dir_path + else: + dir_path = os.path.dirname(dir_path) + + raise RuntimeError("Could not find repo root") + + +def initialize_logger(rank, logger_file_name): + logger = logging.getLogger() + logger.setLevel(logging.INFO) + fh = logging.FileHandler(logger_file_name + f"_{rank}.log", mode="w") + fh.setLevel(logging.INFO) + logger.addHandler(fh) + return logger + + +# This is required for env initialization since we use mpirun +def initialize_distributed_env(logger_file_name, rank=0, world_size=1, port=29500): + local_rank = int( + os.environ.get("OMPI_COMM_WORLD_LOCAL_RANK", rank % torch.cuda.device_count()) + ) + world_size = int(os.environ.get("OMPI_COMM_WORLD_SIZE", world_size)) + + # Set up environment variable to run with mpirun + os.environ["RANK"] = str(local_rank) + os.environ["WORLD_SIZE"] = str(world_size) + os.environ["MASTER_ADDR"] = "127.0.0.1" + os.environ["MASTER_PORT"] = str(port) + os.environ["TRTLLM_PLUGINS_PATH"] = ( + find_repo_root() + "/lib/libnvinfer_plugin_tensorrt_llm.so" + ) + + # Necessary to assign a device to each rank. + torch.cuda.set_device(local_rank) + + # We use nccl backend + dist.init_process_group("nccl") + + # set a manual seed for reproducibility + torch.manual_seed(1111) + + device_mesh = init_device_mesh(device_type="cuda", mesh_shape=(world_size,)) + rank = device_mesh.get_rank() + assert rank == local_rank + logger = initialize_logger(rank, logger_file_name) + device_id = ( + rank % torch.cuda.device_count() + ) # Ensure each rank gets a unique device + torch.cuda.set_device(device_id) + + return device_mesh, world_size, rank, logger diff --git a/examples/distributed_inference/tensor_parallel_llama3.py b/examples/distributed_inference/tensor_parallel_llama3.py index fc03a64386..998c378be2 100644 --- a/examples/distributed_inference/tensor_parallel_llama3.py +++ b/examples/distributed_inference/tensor_parallel_llama3.py @@ -5,27 +5,25 @@ import time import torch -import torch_tensorrt from llama3_model import ModelArgs, ParallelTransformer +from tensor_parallel_initialize_dist import initialize_distributed_env from torch.distributed._composable.fsdp import MixedPrecisionPolicy from torch.distributed._composable.fsdp.fully_shard import fully_shard from torch.distributed._tensor import Replicate, Shard from torch.distributed.algorithms._checkpoint.checkpoint_wrapper import ( checkpoint_wrapper, ) -from torch.distributed.device_mesh import DeviceMesh, init_device_mesh -_rank = int(os.environ["RANK"]) -_world_size = int(os.environ["WORLD_SIZE"]) -tp_size = 2 - -logger = logging.getLogger() -logger.setLevel(logging.INFO) -fh = logging.FileHandler(f"./tensor_parallel_log_{_rank}.log", mode="w") -fh.setLevel(logging.INFO) -logger.addHandler(fh) +device_mesh, _world_size, _rank, logger = initialize_distributed_env( + "./tensor_parallel_llama3" +) +# Import should be after initialization of the TRT-LLM plugin .so path +import tensorrt_llm -tp_mesh = init_device_mesh(device_type="cuda", mesh_shape=(_world_size,)) +logger.info(f"Starting PyTorch TP example on rank {_rank}.") +assert ( + _world_size % 2 == 0 +), f"TP examples require even number of GPUs, but got {_world_size} gpus" model_args = ModelArgs( vocab_size=32000, @@ -38,7 +36,7 @@ ) with torch.no_grad(): - model = ParallelTransformer(model_args, tp_mesh) + model = ParallelTransformer(model_args, device_mesh) torch.manual_seed(0) inp = torch.randint(32000, (8, 256), device="cuda") python_result = model(inp) @@ -53,7 +51,7 @@ "use_python_runtime": True, "workspace_size": 1 << 33, "debug": False, - "timing_cache_path": "/opt/file/cache/timing_cache_llama.bin", + "use_aot_joint_export": False, }, dynamic=False, ) diff --git a/examples/distributed_inference/tensor_parallel_simple_example.py b/examples/distributed_inference/tensor_parallel_simple_example.py index 470487a751..837648fdb4 100755 --- a/examples/distributed_inference/tensor_parallel_simple_example.py +++ b/examples/distributed_inference/tensor_parallel_simple_example.py @@ -1,18 +1,22 @@ -import os -import sys import time +import tensorrt as trt import torch import torch.nn as nn import torch_tensorrt +from tensor_parallel_initialize_dist import initialize_distributed_env from torch.distributed._tensor import Shard -from torch.distributed._tensor.device_mesh import init_device_mesh from torch.distributed.tensor.parallel import ( ColwiseParallel, RowwiseParallel, parallelize_module, ) +device_mesh, _world_size, _rank, logger = initialize_distributed_env( + "./tensor_parallel_simple_example" +) +import tensorrt_llm + """ This example copies some code from https://github.com/pytorch/examples/blob/main/distributed/tensor_parallelism/tensor_parallel_example.py """ @@ -36,14 +40,7 @@ def forward(self, x): return x -# create a device mesh based on the given world_size. -_world_size = int(os.environ["WORLD_SIZE"]) - -device_mesh = init_device_mesh(device_type="cuda", mesh_shape=(_world_size,)) -_rank = device_mesh.get_rank() - - -print(f"Starting PyTorch TP example on rank {_rank}.") +logger.info(f"Starting PyTorch TP example on rank {_rank}.") assert ( _world_size % 2 == 0 ), f"TP examples require even number of GPUs, but got {_world_size} gpus" @@ -78,6 +75,7 @@ def forward(self, x): "enabled_precisions": {torch.float32, torch.float16}, "use_python_runtime": True, "min_block_size": 1, + "use_aot_joint_export": False, }, dynamic=False, ) @@ -91,9 +89,9 @@ def forward(self, x): output = tp_model(inp) end = time.time() if i == 0: - print(f"Compilation time is {end-start}") + logger.info(f"Compilation time is {end-start}") assert ( python_result - output ).std() < 0.01, "Compilation result is not correct." elif _rank == 0: - print(f"Inference time is {end-start}") + logger.info(f"Inference time is {end-start}") diff --git a/noxfile.py b/noxfile.py index 73d0dca854..9a5c3263bb 100644 --- a/noxfile.py +++ b/noxfile.py @@ -34,7 +34,7 @@ # Set epochs to train VGG model for accuracy tests EPOCHS = 25 -SUPPORTED_PYTHON_VERSIONS = ["3.7", "3.8", "3.9", "3.10"] +SUPPORTED_PYTHON_VERSIONS = ["3.9", "3.10", "3.11", "3.12"] nox.options.sessions = [ "l0_api_tests-" + "{}.{}".format(sys.version_info.major, sys.version_info.minor) diff --git a/packaging/env_vars.txt b/packaging/env_vars.txt index bfd05b7bf6..44a2350e0f 100644 --- a/packaging/env_vars.txt +++ b/packaging/env_vars.txt @@ -1,7 +1,3 @@ export CI_BUILD="1" export RELEASE="1" -if [ $CU_VERSION == "cu126" ]; then - export USE_CXX11_ABI="1" -else - export USE_CXX11_ABI="0" -fi \ No newline at end of file +export USE_CXX11_ABI="1" \ No newline at end of file diff --git a/packaging/smoke_test_script.sh b/packaging/smoke_test_script.sh index 19d9d717a4..ef584fef51 100644 --- a/packaging/smoke_test_script.sh +++ b/packaging/smoke_test_script.sh @@ -2,5 +2,5 @@ # The issue was smoke test installs the built torch_tensorrt wheel file and checks `import torch_tensorrt; print(torch_tensorrt.__version__)` # Since tensorrt cannot be pip installable in CI, the smoke test will fail. # One way we tried to handle it is manually install tensorrt wheel while by extracting from the tarball. -# However, the TensorRT-10.3.0.26/lib path doesn't seem to show up in LD_LIBRARY_PATH even if we explicitly set it. +# However, the TensorRT-10.7.0.23/lib path doesn't seem to show up in LD_LIBRARY_PATH even if we explicitly set it. # TODO: Implement a custom smoke_test script to verify torch_tensorrt installation. \ No newline at end of file diff --git a/py/ci/Dockerfile.ci b/py/ci/Dockerfile.ci index 823c8bb7a1..288c61029d 100644 --- a/py/ci/Dockerfile.ci +++ b/py/ci/Dockerfile.ci @@ -3,13 +3,14 @@ FROM pytorch/manylinux2_28-builder:cuda12.6 RUN yum install -y ninja-build # download TensorRT tarball -RUN wget -q https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.3.0/tars/TensorRT-10.3.0.26.Linux.x86_64-gnu.cuda-12.5.tar.gz \ -&& gunzip TensorRT-10.3.0.26.Linux.x86_64-gnu.cuda-12.5.tar.gz \ -&& tar -xvf TensorRT-10.3.0.26.Linux.x86_64-gnu.cuda-12.5.tar \ -&& rm TensorRT-10.3.0.26.Linux.x86_64-gnu.cuda-12.5.tar +RUN wget -q https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.7.0/tars/TensorRT-10.7.0.23.Linux.x86_64-gnu.cuda-12.6.tar.gz \ +&& gunzip TensorRT-10.7.0.23.Linux.x86_64-gnu.cuda-12.6.tar.gz \ +&& tar -xvf TensorRT-10.7.0.23.Linux.x86_64-gnu.cuda-12.6.tar \ +&& rm TensorRT-10.7.0.23.Linux.x86_64-gnu.cuda-12.6.tar -ENV TENSORRT_DIR=/TensorRT-10.3.0.26 -ENV TENSORRT_VERSION=10.3.0 +ENV TENSORRT_DIR=/TensorRT-10.7.0.23 +ENV TENSORRT_VERSION=10.7.0 +ENV USE_CXX11_ABI=1 RUN wget https://github.com/bazelbuild/bazelisk/releases/download/v1.17.0/bazelisk-linux-amd64 \ && mv bazelisk-linux-amd64 /usr/bin/bazel \ diff --git a/py/requirements.txt b/py/requirements.txt index 361afab365..5644656330 100644 --- a/py/requirements.txt +++ b/py/requirements.txt @@ -2,7 +2,7 @@ numpy packaging pybind11==2.6.2 --extra-index-url https://download.pytorch.org/whl/nightly/cu124 -torch>=2.6.0.dev,<2.7.0 -torchvision>=0.20.0.dev,<0.21.0 +torch>=2.7.0.dev,<2.8.0 +torchvision>=0.22.0.dev,<0.23.0 --extra-index-url https://pypi.ngc.nvidia.com pyyaml diff --git a/py/torch_tensorrt/dynamo/_defaults.py b/py/torch_tensorrt/dynamo/_defaults.py index 76630a75a5..18932e6cd0 100644 --- a/py/torch_tensorrt/dynamo/_defaults.py +++ b/py/torch_tensorrt/dynamo/_defaults.py @@ -46,6 +46,7 @@ IMMUTABLE_WEIGHTS = True ENABLE_WEIGHT_STREAMING = False ENABLE_CROSS_COMPILE_FOR_WINDOWS = False +USE_AOT_JOINT_EXPORT = True def default_device() -> Device: diff --git a/py/torch_tensorrt/dynamo/_settings.py b/py/torch_tensorrt/dynamo/_settings.py index 7a22663af3..05fb5ce094 100644 --- a/py/torch_tensorrt/dynamo/_settings.py +++ b/py/torch_tensorrt/dynamo/_settings.py @@ -33,6 +33,7 @@ STRIP_ENGINE_WEIGHTS, TIMING_CACHE_PATH, TRUNCATE_DOUBLE, + USE_AOT_JOINT_EXPORT, USE_EXPLICIT_TYPING, USE_FAST_PARTITIONER, USE_FP32_ACC, @@ -91,6 +92,7 @@ class CompilationSettings: enable_weight_streaming (bool): Enable weight streaming. enable_cross_compile_for_windows (bool): By default this is False means TensorRT engines can only be executed on the same platform where they were built. True will enable cross-platform compatibility which allows the engine to be built on Linux and run on Windows + use_aot_joint_export (bool): Use aot_export_joint_simple, else wrap backend with AOT_autograd, required for distributed tensors """ enabled_precisions: Set[dtype] = field(default_factory=lambda: ENABLED_PRECISIONS) @@ -131,6 +133,7 @@ class CompilationSettings: immutable_weights: bool = IMMUTABLE_WEIGHTS enable_weight_streaming: bool = ENABLE_WEIGHT_STREAMING enable_cross_compile_for_windows: bool = ENABLE_CROSS_COMPILE_FOR_WINDOWS + use_aot_joint_export: bool = USE_AOT_JOINT_EXPORT _SETTINGS_TO_BE_ENGINE_INVARIANT = ( diff --git a/py/torch_tensorrt/dynamo/backend/backends.py b/py/torch_tensorrt/dynamo/backend/backends.py index 59bd7d011d..ef04745562 100644 --- a/py/torch_tensorrt/dynamo/backend/backends.py +++ b/py/torch_tensorrt/dynamo/backend/backends.py @@ -1,17 +1,20 @@ from __future__ import annotations +import functools import logging import unittest from typing import Any, Callable, Sequence import torch import torch._dynamo as td +from torch._dynamo.backends.common import aot_autograd from torch._dynamo.utils import detect_fake_mode from torch._functorch.aot_autograd import aot_export_joint_simple from torch_tensorrt.dynamo import CompilationSettings from torch_tensorrt.dynamo._compiler import compile_module from torch_tensorrt.dynamo.lowering import ( get_decompositions, + modify_reshape_complex_nodes, post_lowering, remove_detach, remove_sym_nodes, @@ -49,7 +52,25 @@ def aot_torch_tensorrt_aten_backend( gm: torch.fx.GraphModule, sample_inputs: Sequence[Any], **kwargs: Any ) -> torch.nn.Module: settings, engine_cache = parse_dynamo_kwargs(kwargs) - return _pretraced_backend(gm, sample_inputs, settings, engine_cache) + if settings.use_aot_joint_export: + return _pretraced_backend(gm, sample_inputs, settings, engine_cache) + logger.debug("Wrapping the backend with aot_autograd\n") + _pretraced_backend_autograd = functools.partial( + _pretraced_backend, settings=settings, engine_cache=engine_cache + ) + settings_aot_autograd = {} + settings_aot_autograd["decompostions"] = get_decompositions( + settings.enable_experimental_decompositions + ) + # This is added since detach lowering leads to alias nodes + # Error - View operation returned a tensor that is the same as the input base tensor + # torch nop_decompositions in torch/_decomp/decompositions.py + if aten.detach in settings_aot_autograd["decompositions"]: + del settings_aot_autograd["decompositions"][aten.detach] + return aot_autograd( + fw_compiler=_pretraced_backend_autograd, + decompositions=get_decompositions(settings.enable_experimental_decompositions), + )(gm, sample_inputs) def _pretraced_backend( @@ -89,15 +110,26 @@ def _pretraced_backend( # Remove detach nodes remove_detach(gm, settings) + complexInputIndices = [] + for i, torch_input in enumerate(torch_inputs): + if torch_inputs[i].dtype == torch.complex64: + complexInputIndices.append(i) + torch_input_real = torch_inputs[i].real + torch_input_imaginary = torch_inputs[i].imag + torch_inputs[i] = torch.stack( + (torch_input_real, torch_input_imaginary), dim=-1 + ) + # Invoke AOTAutograd to translate operators to aten - gm = aot_export_joint_simple( - gm, - sample_inputs, - trace_joint=False, - decompositions=get_decompositions( - settings.enable_experimental_decompositions - ), - ) + if settings.use_aot_joint_export: + gm = aot_export_joint_simple( + gm, + sample_inputs, + trace_joint=False, + decompositions=get_decompositions( + settings.enable_experimental_decompositions + ), + ) logger.debug("Post-AOT Autograd graph:\n" + str(gm.graph)) @@ -105,6 +137,12 @@ def _pretraced_backend( logger.debug("Lowered Input graph:\n " + str(gm.graph)) + if complexInputIndices: + modify_reshape_complex_nodes(gm, complexInputIndices) + logger.debug( + "Input graph after modifying complex nodes:\n " + str(gm.graph) + ) + torchtrt_inputs = prepare_inputs( torch_inputs, disable_memory_format_check=True ) diff --git a/py/torch_tensorrt/dynamo/conversion/__init__.py b/py/torch_tensorrt/dynamo/conversion/__init__.py index 26a428ed8d..f41540a0cf 100644 --- a/py/torch_tensorrt/dynamo/conversion/__init__.py +++ b/py/torch_tensorrt/dynamo/conversion/__init__.py @@ -1,4 +1,10 @@ -from . import aten_ops_converters, ops_evaluators, plugins, prims_ops_converters +from . import ( + aten_ops_converters, + custom_ops_converters, + ops_evaluators, + plugins, + prims_ops_converters, +) from ._conversion import convert_module, interpret_module_to_result from ._ConversionContext import ConversionContext from ._ConverterRegistry import * # noqa: F403 diff --git a/py/torch_tensorrt/dynamo/conversion/aten_ops_converters.py b/py/torch_tensorrt/dynamo/conversion/aten_ops_converters.py index 16d20d3c76..5254c6a0ac 100644 --- a/py/torch_tensorrt/dynamo/conversion/aten_ops_converters.py +++ b/py/torch_tensorrt/dynamo/conversion/aten_ops_converters.py @@ -2,7 +2,7 @@ import logging import operator -from typing import Any, Callable, Dict, Optional, Sequence, Tuple, Union +from typing import Callable, Dict, Optional, Sequence, Tuple, Union import numpy as np import torch @@ -16,6 +16,7 @@ has_static_shapes_in_args, ) from torch_tensorrt.dynamo.conversion.converter_utils import ( + args_bounds_check, enforce_tensor_types, get_positive_dim, is_only_operator_on_placeholder, @@ -25,12 +26,6 @@ _LOGGER: logging.Logger = logging.getLogger(__name__) -def args_bounds_check( - args: Tuple[Argument, ...], i: int, replacement: Optional[Any] = None -) -> Any: - return args[i] if len(args) > i and args[i] is not None else replacement - - def get_ir(target: Target) -> SourceIR: target_module = getattr(target, "__module__", "None") if any( diff --git a/py/torch_tensorrt/dynamo/conversion/converter_utils.py b/py/torch_tensorrt/dynamo/conversion/converter_utils.py index 937ed27fe4..f4bb4877cc 100644 --- a/py/torch_tensorrt/dynamo/conversion/converter_utils.py +++ b/py/torch_tensorrt/dynamo/conversion/converter_utils.py @@ -1,10 +1,11 @@ import collections +import ctypes import functools import logging +import os from typing import Any, Callable, Dict, List, Optional, Sequence, Tuple, Union, overload import numpy as np -import tensorrt as trt import torch import torch_tensorrt.dynamo.conversion.impl as impl from torch.fx.node import Argument, Target @@ -18,6 +19,8 @@ DynamoConverterImplSignature, ) +import tensorrt as trt + from ..types import Shape, TRTDataType, TRTLayer, TRTTensor _LOGGER: logging.Logger = logging.getLogger(__name__) @@ -913,3 +916,72 @@ def set_layer_name( else f"{source_ir}_ops.{target.__name__}" ) layer.name = f"[{layer.type.name}]-[{target_name}]-[{name}]" + + +def args_bounds_check( + args: Tuple[Argument, ...], i: int, replacement: Optional[Any] = None +) -> Any: + return args[i] if len(args) > i and args[i] is not None else replacement + + +def load_tensorrt_llm() -> bool: + """ + Attempts to load the TensorRT-LLM plugin and initialize it. + + Returns: + bool: True if the plugin was successfully loaded and initialized, False otherwise. + """ + try: + import tensorrt_llm as trt_llm # noqa: F401 + + _LOGGER.info("TensorRT-LLM successfully imported") + return True + except (ImportError, AssertionError) as e_import_error: + # Check for environment variable for the plugin library path + plugin_lib_path = os.environ.get("TRTLLM_PLUGINS_PATH") + if not plugin_lib_path: + _LOGGER.warning( + "TensorRT-LLM is not installed. Please install TensorRT-LLM or set TRTLLM_PLUGINS_PATH to the directory containing libnvinfer_plugin_tensorrt_llm.so to use converters for torch.distributed ops", + ) + return False + + _LOGGER.info(f"TensorRT-LLM Plugin lib path found: {plugin_lib_path}") + try: + # Load the shared library + handle = ctypes.CDLL(plugin_lib_path) + _LOGGER.info(f"Successfully loaded plugin library: {plugin_lib_path}") + except OSError as e_os_error: + _LOGGER.error( + f"Failed to load libnvinfer_plugin_tensorrt_llm.so from {plugin_lib_path}" + f"Ensure the path is correct and the library is compatible", + exc_info=e_os_error, + ) + return False + + try: + # Configure plugin initialization arguments + handle.initTrtLlmPlugins.argtypes = [ctypes.c_void_p, ctypes.c_char_p] + handle.initTrtLlmPlugins.restype = ctypes.c_bool + except AttributeError as e_plugin_unavailable: + _LOGGER.warning( + "Unable to initialize the TensorRT-LLM plugin library", + exc_info=e_plugin_unavailable, + ) + return False + + try: + # Initialize the plugin + TRT_LLM_PLUGIN_NAMESPACE = "tensorrt_llm" + if handle.initTrtLlmPlugins(None, TRT_LLM_PLUGIN_NAMESPACE.encode("utf-8")): + _LOGGER.info("TensorRT-LLM plugin successfully initialized") + return True + else: + _LOGGER.warning("TensorRT-LLM plugin library failed in initialization") + return False + except Exception as e_initialization_error: + _LOGGER.warning( + "Exception occurred during TensorRT-LLM plugin library initialization", + exc_info=e_initialization_error, + ) + return False + return False diff --git a/py/torch_tensorrt/dynamo/conversion/custom_ops_converters.py b/py/torch_tensorrt/dynamo/conversion/custom_ops_converters.py new file mode 100644 index 0000000000..17850fabce --- /dev/null +++ b/py/torch_tensorrt/dynamo/conversion/custom_ops_converters.py @@ -0,0 +1,61 @@ +# mypy: disallow-untyped-decorators=False + +import logging +from typing import Dict, Sequence, Tuple, Union + +from torch.fx.node import Argument, Target +from torch_tensorrt.dynamo._SourceIR import SourceIR +from torch_tensorrt.dynamo.conversion import impl +from torch_tensorrt.dynamo.conversion._ConversionContext import ConversionContext +from torch_tensorrt.dynamo.conversion._ConverterRegistry import ( + dynamo_tensorrt_converter, +) +from torch_tensorrt.dynamo.conversion.converter_utils import load_tensorrt_llm +from torch_tensorrt.dynamo.lowering.passes.fuse_distributed_ops import ( + tensorrt_fused_nccl_all_gather_op, + tensorrt_fused_nccl_reduce_scatter_op, +) + +import tensorrt as trt + +_LOGGER: logging.Logger = logging.getLogger(__name__) + +if load_tensorrt_llm(): + + @dynamo_tensorrt_converter(tensorrt_fused_nccl_all_gather_op) + def fused_nccl_gather( + ctx: ConversionContext, + target: Target, + args: Tuple[Argument, ...], + kwargs: Dict[str, Argument], + name: str, + ) -> Union[trt.ITensor, Sequence[trt.ITensor]]: + return impl.distributed.nccl_gather( + ctx, + target, + SourceIR.ATEN, + name, + [args[0]], + ) + + @dynamo_tensorrt_converter(tensorrt_fused_nccl_reduce_scatter_op) + def fused_nccl_reduce_scatter( + ctx: ConversionContext, + target: Target, + args: Tuple[Argument, ...], + kwargs: Dict[str, Argument], + name: str, + ) -> Union[trt.ITensor, Sequence[trt.ITensor]]: + return impl.distributed.nccl_reduce_scatter( + ctx, + target, + SourceIR.ATEN, + name, + [args[0]], + ) + + breakpoint() +else: + _LOGGER.debug( + "Did not load torch.distributed converters since TensorRT-LLM is not available" + ) diff --git a/py/torch_tensorrt/dynamo/conversion/impl/__init__.py b/py/torch_tensorrt/dynamo/conversion/impl/__init__.py index c1187f0dd9..75f7492591 100644 --- a/py/torch_tensorrt/dynamo/conversion/impl/__init__.py +++ b/py/torch_tensorrt/dynamo/conversion/impl/__init__.py @@ -13,6 +13,7 @@ full, grid, matmul, + nccl_ops, normalization, pad, permutation, diff --git a/py/torch_tensorrt/dynamo/conversion/impl/elementwise/base.py b/py/torch_tensorrt/dynamo/conversion/impl/elementwise/base.py index 3ce1714a7d..4a4b33abea 100644 --- a/py/torch_tensorrt/dynamo/conversion/impl/elementwise/base.py +++ b/py/torch_tensorrt/dynamo/conversion/impl/elementwise/base.py @@ -3,7 +3,6 @@ from typing import Any, Callable, Optional, Union import numpy as np -import tensorrt as trt import torch from torch.fx.node import Target from torch_tensorrt import _enums @@ -14,12 +13,13 @@ broadcast_to_same_shape, cast_trt_tensor, get_trt_tensor, - broadcast, has_dynamic_shape, set_layer_name, ) from torch_tensorrt.dynamo.types import TRTElementWiseOp, TRTTensor +import tensorrt as trt + def get_python_op_from_trt_elementwise_op( trt_op: TRTElementWiseOp, diff --git a/py/torch_tensorrt/dynamo/conversion/impl/nccl_ops.py b/py/torch_tensorrt/dynamo/conversion/impl/nccl_ops.py new file mode 100644 index 0000000000..013268f803 --- /dev/null +++ b/py/torch_tensorrt/dynamo/conversion/impl/nccl_ops.py @@ -0,0 +1,120 @@ +import os +from enum import IntEnum, IntFlag, auto +from typing import Optional, Tuple, Union + +import numpy as np +from torch.fx.node import Argument, Target +from torch_tensorrt.dynamo.conversion._ConversionContext import ConversionContext +from torch_tensorrt.fx.converters.converter_utils import SourceIR, set_layer_name + +import tensorrt as trt + + +# class for AllReduce +class AllReduceStrategy(IntEnum): + """Warning: actual definition is in kernels/customAllReduceKernels.h. + + They must be kept in sync. + """ + + NCCL = 0 + ONESHOT = 1 + TWOSHOT = 2 + AUTO = 3 + + +class AllReduceConfig(IntFlag): + """Warning: actual definition is in kernels/customAllReduceKernels.h. + + They must be kept in sync + """ + + USE_MEMCPY = auto() + PUSH_MODE = auto() + + +def nccl_gather( + ctx: ConversionContext, + target: Union[Target, str], + source_ir: Optional[SourceIR], + name: str, + plug_inputs: Tuple[Argument, ...], +) -> trt.ITensor: + allgather_plg_creator = trt.get_plugin_registry().get_plugin_creator( + "AllGather", "1", "tensorrt_llm" + ) + assert allgather_plg_creator is not None + _world_size = os.environ.get("WORLD_SIZE") + if _world_size is not None: + world_size = int(_world_size) + else: + raise RuntimeError( + "The WORLD_SIZE env variable is not set in distributed environment" + ) + group = list(range(world_size)) + group = trt.PluginField( + "group", np.array(group, dtype=np.int32), trt.PluginFieldType.INT32 + ) + p_dtype = trt.float32 + pf_type = trt.PluginField( + "type_id", np.array([int(p_dtype)], np.int32), trt.PluginFieldType.INT32 + ) + pfc = trt.PluginFieldCollection([group, pf_type]) + allgather = allgather_plg_creator.create_plugin("allgather", pfc) + layer = ctx.net.add_plugin_v2(plug_inputs, allgather) + set_layer_name(layer, target, name, source_ir) + return layer.get_output(0) + + +def nccl_reduce_scatter( + ctx: ConversionContext, + target: Union[Target, str], + source_ir: Optional[SourceIR], + name: str, + plug_inputs: Tuple[Argument, ...], +) -> trt.ITensor: + allreduce_plg_creator = trt.get_plugin_registry().get_plugin_creator( + "ReduceScatter", "1", "tensorrt_llm" + ) + + assert allreduce_plg_creator is not None + + counter = 0 + strategy = AllReduceStrategy.NCCL + config = AllReduceConfig(0) + _world_size = os.environ.get("WORLD_SIZE") + if _world_size is not None: + world_size = int(_world_size) + else: + raise RuntimeError( + "The WORLD_SIZE env variable is not set in distributed environment" + ) + group = list(range(world_size)) + group = trt.PluginField( + "group", np.array(group, dtype=np.int32), trt.PluginFieldType.INT32 + ) + + p_dtype = trt.float16 + pf_dtype = trt.PluginField( + "type_id", np.array([int(p_dtype)], np.int32), trt.PluginFieldType.INT32 + ) + pfc = [group, pf_dtype] + p_strategy = trt.PluginField( + "strategy", np.array([int(strategy)], np.int8), trt.PluginFieldType.INT8 + ) + pfc.append(p_strategy) + p_config = trt.PluginField( + "config", np.array([int(config)], np.int8), trt.PluginFieldType.INT8 + ) + pfc.append(p_config) + p_counter = trt.PluginField( + "counter", np.array([counter], np.int32), trt.PluginFieldType.INT32 + ) + pfc.append(p_counter) + + pfc = trt.PluginFieldCollection(pfc) + ar_plug = allreduce_plg_creator.create_plugin("allreduce", pfc) + + layer = ctx.net.add_plugin_v2(plug_inputs, ar_plug) + set_layer_name(layer, target, name, source_ir) + return layer.get_output(0) diff --git a/py/torch_tensorrt/dynamo/conversion/impl/select.py b/py/torch_tensorrt/dynamo/conversion/impl/select.py index 0d55c5f014..dc5458d2c8 100644 --- a/py/torch_tensorrt/dynamo/conversion/impl/select.py +++ b/py/torch_tensorrt/dynamo/conversion/impl/select.py @@ -2,7 +2,6 @@ from typing import Optional, Sequence, Union import numpy as np -import tensorrt as trt import torch from torch.fx.node import Target from torch_tensorrt.dynamo._SourceIR import SourceIR @@ -16,7 +15,6 @@ to_numpy, ) from torch_tensorrt.dynamo.conversion.impl.elementwise import convert_binary_elementwise -from torch_tensorrt.dynamo.conversion.impl.shape import get_shape_with_dynamic_shape from torch_tensorrt.dynamo.conversion.impl.shape import shape as get_shape from torch_tensorrt.dynamo.utils import DYNAMIC_DIM from torch_tensorrt.fx.converters.converter_utils import ( @@ -25,6 +23,8 @@ ) from torch_tensorrt.fx.types import TRTTensor +import tensorrt as trt + _LOGGER: logging.Logger = logging.getLogger(__name__) diff --git a/py/torch_tensorrt/dynamo/lowering/passes/__init__.py b/py/torch_tensorrt/dynamo/lowering/passes/__init__.py index c0e2803e60..716c6505fe 100644 --- a/py/torch_tensorrt/dynamo/lowering/passes/__init__.py +++ b/py/torch_tensorrt/dynamo/lowering/passes/__init__.py @@ -1,3 +1,4 @@ from ._aten_lowering_pass import * +from ._modify_reshape_complex_nodes import modify_reshape_complex_nodes from .remove_sym_nodes import remove_sym_nodes from .repair_input_aliasing import repair_input_aliasing diff --git a/py/torch_tensorrt/dynamo/lowering/passes/_aten_lowering_pass.py b/py/torch_tensorrt/dynamo/lowering/passes/_aten_lowering_pass.py index 661c76d3b6..f24fd8ec21 100644 --- a/py/torch_tensorrt/dynamo/lowering/passes/_aten_lowering_pass.py +++ b/py/torch_tensorrt/dynamo/lowering/passes/_aten_lowering_pass.py @@ -6,6 +6,7 @@ from .accumulate_fp32_matmul import accumulate_fp32_matmul from .constant_folding import constant_fold +from .fuse_distributed_ops import fuse_distributed_ops from .fuse_prims_broadcast import fuse_prims_broadcast from .lower_scaled_dot_product_attention import lower_scaled_dot_product_attention from .pass_manager import DynamoPassManager @@ -22,6 +23,7 @@ constant_fold, repair_input_as_output, fuse_prims_broadcast, + fuse_distributed_ops, replace_max_pool_with_indices, lower_scaled_dot_product_attention, view_to_reshape, diff --git a/py/torch_tensorrt/dynamo/lowering/passes/_modify_reshape_complex_nodes.py b/py/torch_tensorrt/dynamo/lowering/passes/_modify_reshape_complex_nodes.py new file mode 100644 index 0000000000..f8ca1f71b9 --- /dev/null +++ b/py/torch_tensorrt/dynamo/lowering/passes/_modify_reshape_complex_nodes.py @@ -0,0 +1,105 @@ +import logging + +import torch + +logger = logging.getLogger(__name__) + +from torch_tensorrt.dynamo.lowering.passes.pass_utils import ( + clean_up_graph_after_modifications, + find_complex_nodes, +) + +from ._replace_complex_placeholder_to_tuple import replace_complex_placeholder_to_tuple + + +def tensorrt_complex_mul(args0, args1): + args0_real, args0_imag = torch.ops.aten.split.Tensor(args0, 1, -1) + args1_real, args1_imag = torch.ops.aten.split.Tensor(args1, 1, -1) + + args0_real = torch.ops.aten.squeeze.dim(args0_real, -1) + args0_imag = torch.ops.aten.squeeze.dim(args0_imag, -1) + args1_real = torch.ops.aten.squeeze.dim(args1_real, -1) + args1_imag = torch.ops.aten.squeeze.dim(args1_imag, -1) + + complex_mul_real = torch.ops.aten.sub( + torch.ops.aten.mul(args0_real, args1_real), + torch.ops.aten.mul(args0_imag, args1_imag), + ) + complex_mul_imag = torch.ops.aten.add( + torch.ops.aten.mul(args0_real, args1_imag), + torch.ops.aten.mul(args0_imag, args1_real), + ) + + return torch.ops.aten.stack((complex_mul_real, complex_mul_imag), -1) + + +def remove_complex_real_view_nodes(gm: torch.fx.GraphModule): + modified_graph = False + nodes_to_remove = [] + for node in gm.graph.nodes: + if "view_as_complex" in node.name or "view_as_real" in node.name: + nodes_to_remove.append(node) + + for node in nodes_to_remove: + input_node = node.args[0] if node.args else None + + for other_node in gm.graph.nodes: + new_args = tuple( + input_node if arg is node else arg for arg in other_node.args + ) + other_node.args = new_args + + gm.graph.erase_node(node) + modified_graph = True + + if modified_graph: + gm = clean_up_graph_after_modifications(gm) + logger.debug( + f"Graph after removing view_as_complex nodes and view_as_real nodes:\n{gm.graph}" + ) + + +def modify_reshape_nodes(gm: torch.fx.GraphModule, complex_nodes): + for node in gm.graph.nodes: + if node in complex_nodes: + # slice and transpose will remain same + if "reshape" in node.name: + new_shape = list(node.args[1]) + [2] + node.args = (node.args[0], tuple(new_shape)) + + +def modify_mul_nodes(gm: torch.fx.GraphModule, complex_nodes): + modified_graph = False + for node in gm.graph.nodes: + if node in complex_nodes: + if "mul" in node.name: + complex_mul_args = (node.args[0], node.args[1]) + with gm.graph.inserting_after(node): + replacement_node = gm.graph.create_node( + op="call_function", + target=tensorrt_complex_mul, + args=complex_mul_args, + ) + node.replace_all_uses_with(replacement_node) + replacement_node.meta.update(node.meta) + modified_graph = True + gm.graph.erase_node(node) + + if modified_graph: + gm = clean_up_graph_after_modifications(gm) + logger.debug( + f"Graph after custom complex mul nodes is applied to the graph:\n{gm.graph}" + ) + + +def modify_complex_nodes(gm: torch.fx.GraphModule, complex_nodes): + modify_reshape_nodes(gm, complex_nodes) + remove_complex_real_view_nodes(gm) + modify_mul_nodes(gm, complex_nodes) + + +def modify_reshape_complex_nodes(gm: torch.fx.GraphModule, complexInputIndices): + complex_nodes = find_complex_nodes(gm) + if complex_nodes: + replace_complex_placeholder_to_tuple(gm, complexInputIndices) + modify_complex_nodes(gm, complex_nodes) diff --git a/py/torch_tensorrt/dynamo/lowering/passes/_replace_complex_placeholder_to_tuple.py b/py/torch_tensorrt/dynamo/lowering/passes/_replace_complex_placeholder_to_tuple.py new file mode 100644 index 0000000000..e2edec3d28 --- /dev/null +++ b/py/torch_tensorrt/dynamo/lowering/passes/_replace_complex_placeholder_to_tuple.py @@ -0,0 +1,112 @@ +import logging +from typing import List, Tuple + +import torch +from torch._subclasses.fake_tensor import FakeTensorMode +from torch.fx.node import _get_qualified_name +from torch_tensorrt.dynamo.conversion.converter_utils import args_bounds_check + +# dead-code elimination, linting, and recompilation for graph, in-place +from torch_tensorrt.dynamo.lowering.passes.pass_utils import ( + clean_up_graph_after_modifications, +) + +logger = logging.getLogger(__name__) + + +def replace_complex_placeholder_to_tuple( + gm: torch.fx.GraphModule, + inputListindices: List[int], +) -> torch.fx.GraphModule: + modified_graph = False + input_arg_list = [f"arg{inputListIndex}_1" for inputListIndex in inputListindices] + for node in gm.graph.nodes: + if node.op == "placeholder" and node.target in input_arg_list: + from torch._subclasses.fake_tensor import FakeTensorMode + + node_shape = node.meta["val"].size() + new_node_shape = node_shape + (2,) + new_node_dtype = None + if node.meta["val"].dtype == torch.complex64: + new_node_dtype = torch.float32 + else: + new_node_dtype = torch.float64 + fake_mode = FakeTensorMode() + + real_tensor = torch.empty(new_node_shape, dtype=new_node_dtype) + with FakeTensorMode() as fake_mode: + new_placeholder_tuple = fake_mode.from_tensor(real_tensor) + node.meta["val"] = new_placeholder_tuple + modified_graph = True + # propagate the meta data change for the downstream ops + # TODO:to check if this is required in all cases + propogate_complex_num_shape_change_till_complex_mul(gm, node, fake_mode) + + # If graph was modified, clean it up + if modified_graph: + gm = clean_up_graph_after_modifications(gm) + logger.debug( + f"Graph after fusing wait_tensor and distributed op tensor:\n{gm.graph}" + ) + + return gm + + +def infer_slice_shape(node: torch.fx.Node) -> Tuple[int, ...]: + input_shape = node.args[0].meta["val"].shape + slice_args = node.args + dim = slice_args[1] + start = slice_args[2] + end = slice_args[3] + step = args_bounds_check(slice_args, 4, replacement=1) + new_shape = list(input_shape) + new_shape[dim] = (end - start + step - 1) // step + return tuple(new_shape) + + +def infer_reshape_shape(node: torch.fx.Node) -> torch.fx.node.Argument: + return node.args[1] + + +shape_inference_funcs = { + "torch.ops.aten.slice.Tensor": infer_slice_shape, + "torch.ops.aten.reshape.default": infer_reshape_shape, +} + + +# Please note this function is for the use case of Llama model +# with complex placeholder->reshape->slice->complex mul +# Hence mul is the terminating op +def propogate_complex_num_shape_change_till_complex_mul( + node: torch.fx.Node, start_node: torch.fx.Node, fake_mode: FakeTensorMode +) -> None: + visited_nodes = set() + stack = [start_node] + while stack: + node = stack.pop() + if node in visited_nodes: + continue + visited_nodes.add(node) + update_node_meta(node, fake_mode) + for user in node.users: + if ( + user.op == "call_function" + and _get_qualified_name(user.target) == "torch.ops.aten.mul.Tensor" + ): + continue + stack.append(user) + + +def update_node_meta(node: torch.fx.Node, fake_mode: FakeTensorMode) -> None: + op_name = node.name + op_target = node.target + + if node.op == "call_function": + op_target = _get_qualified_name(node.target) + + if op_target in shape_inference_funcs: + new_shape = shape_inference_funcs[op_target](node) + real_tensor = torch.empty(new_shape, dtype=node.meta["val"].dtype) + node.meta["val"] = fake_mode.from_tensor(real_tensor) + else: + print("No shape for the inference function", {op_name}) diff --git a/py/torch_tensorrt/dynamo/lowering/passes/fuse_distributed_ops.py b/py/torch_tensorrt/dynamo/lowering/passes/fuse_distributed_ops.py new file mode 100644 index 0000000000..f709f177d6 --- /dev/null +++ b/py/torch_tensorrt/dynamo/lowering/passes/fuse_distributed_ops.py @@ -0,0 +1,80 @@ +import logging +from typing import Any + +import torch +from torch_tensorrt.dynamo._settings import CompilationSettings + +# dead-code elimination, linting, and recompilation for graph, in-place +from torch_tensorrt.dynamo.lowering.passes.pass_utils import ( + clean_up_graph_after_modifications, +) + +logger = logging.getLogger(__name__) + + +# TODO: @apbose make these actual torch custom ops, should allow aot_export +def tensorrt_fused_nccl_all_gather_op( + inp: Any, group_size: int, group_name: str +) -> torch.Tensor: + return torch.ops._c10d_functional.wait_tensor.default( + torch.ops._c10d_functional.all_gather_into_tensor.default( + inp, group_size, group_name + ) + ) + + +def tensorrt_fused_nccl_reduce_scatter_op( + inp: Any, reduce_op: str, group_size: int, group_name: str +) -> torch.Tensor: + return torch.ops._c10d_functional.wait_tensor.default( + torch.ops._c10d_functional.reduce_scatter_tensor.default( + inp, reduce_op, group_size, group_name + ) + ) + + +def fuse_distributed_ops( + gm: torch.fx.GraphModule, settings: CompilationSettings +) -> torch.fx.GraphModule: + modified_graph = False + for node in gm.graph.nodes: + if ( + node.target + in ( + torch.ops._c10d_functional.all_gather_into_tensor.default, + torch.ops._c10d_functional.reduce_scatter_tensor.default, + ) + and len(node.users) == 1 + and list(node.users)[0].target + == torch.ops._c10d_functional.wait_tensor.default + ): + wait_tensor_node = list(node.users)[0] + fused_op = None + if node.target == torch.ops._c10d_functional.all_gather_into_tensor.default: + with gm.graph.inserting_after(wait_tensor_node): + fused_node = gm.graph.create_node( + op="call_function", + target=tensorrt_fused_nccl_all_gather_op, # Define your custom fused function + args=(node.args[0], node.args[1], node.args[2]), + ) + else: + fused_node = gm.graph.create_node( + op="call_function", + target=tensorrt_fused_nccl_reduce_scatter_op, # Define your custom fused function + args=(node.args[0], node.args[1], node.args[2], node.args[3]), + ) + + wait_tensor_node.replace_all_uses_with(fused_node) + fused_node.meta.update(node.meta) + modified_graph = True + gm.graph.erase_node(wait_tensor_node) + gm.graph.erase_node(node) + + # If graph was modified, clean it up + if modified_graph: + gm = clean_up_graph_after_modifications(gm) + logger.debug( + f"Graph after fusing wait_tensor and distributed op tensor:\n{gm.graph}" + ) + + return gm diff --git a/py/torch_tensorrt/dynamo/lowering/passes/pass_utils.py b/py/torch_tensorrt/dynamo/lowering/passes/pass_utils.py index 31a55099c2..1736a234a2 100644 --- a/py/torch_tensorrt/dynamo/lowering/passes/pass_utils.py +++ b/py/torch_tensorrt/dynamo/lowering/passes/pass_utils.py @@ -29,3 +29,42 @@ def get_tensor_placeholders( ] return placeholders + + +def find_complex_nodes(gm: torch.fx.GraphModule): + complex_nodes = [] + complexNodes = {} + for node in gm.graph.nodes: + if is_node_complex(node, complexNodes): + complex_nodes.append(node) + return complex_nodes + + +def is_node_complex(node: torch.fx.Node, complexNodes): + if not isinstance(node, torch.fx.Node): + return False + if node.name in complexNodes: + return True + if node.op == "call_function" and node.args is not None: + for arg in node.args: + if isinstance(arg, int): + continue + elif isinstance(arg, (list, tuple)): + for eachNode in arg: + if is_node_complex(eachNode, complexNodes): + complexNodes[node.name] = True + return True + + elif hasattr(arg, "meta") and "val" in arg.meta: + if isinstance(arg.meta["val"], (list, tuple)): + for eachFakeTensorMeta in arg.meta["val"]: + if eachFakeTensorMeta.dtype in ( + torch.complex64, + torch.complex128, + ): + complexNodes[node.name] = True + return True + elif arg.meta["val"].dtype in (torch.complex64, torch.complex128): + complexNodes[node.name] = True + return True + return False diff --git a/pyproject.toml b/pyproject.toml index 6cc3412849..7ec6c7469a 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -8,8 +8,8 @@ requires = [ "cffi>=1.15.1", "typing-extensions>=4.7.0", "future>=0.18.3", - "tensorrt-cu12>=10.6.0,<10.8.0", - "torch>=2.6.0.dev,<2.7.0", + "tensorrt-cu12>=10.7.0,<10.8.0", + "torch>=2.7.0.dev,<2.8.0", "pybind11==2.6.2", "numpy", ] @@ -54,10 +54,10 @@ keywords = [ "inference", ] dependencies = [ - "torch>=2.6.0.dev,<2.7.0", - "tensorrt-cu12>=10.6.0,<10.8.0", - "tensorrt-cu12-bindings>=10.6.0,<10.8.0", - "tensorrt-cu12-libs>=10.6.0,<10.8.0", + "torch>=2.7.0.dev,<2.8.0", + "tensorrt-cu12>=10.7.0,<10.8.0", + "tensorrt-cu12-bindings>=10.7.0,<10.8.0", + "tensorrt-cu12-libs>=10.7.0,<10.8.0", "packaging>=23", "numpy", "typing-extensions>=4.7.0", @@ -85,9 +85,10 @@ dev = [ torchvision = [ "torchvision", ] #Leaving torchvisions dependency unconstrained so uv can just install something that should work for the torch we have. TV's on PyT makes it hard to put version constrains in -quantization = ["nvidia-modelopt[deploy,hf,torch]~=0.17.0"] +quantization = ["nvidia-modelopt[deploy,hf,torch]>=0.17.0"] monitoring-tools = ["rich>=13.7.1"] jupyter = ["rich[jupyter]>=13.7.1"] +distributed = ["tensorrt-llm>=0.16.0"] [project.urls] Homepage = "https://pytorch.org/tensorrt" @@ -101,17 +102,12 @@ include-package-data = false [tool.uv] environments = ["sys_platform == 'linux'", "sys_platform == 'windows'"] - prerelease = "if-necessary-or-explicit" - +index-strategy = "unsafe-best-match" # Needed for TRT-LLM [tool.uv.sources] -torch = [ - { index = "pytorch-nightly-cu126"}, -] -torchvision = [ - { index = "pytorch-nightly-cu126"}, -] +torch = [{ index = "pytorch-nightly-cu126" }] +torchvision = [{ index = "pytorch-nightly-cu126" }] [[tool.uv.index]] name = "pytorch-nightly-cu126" diff --git a/toolchains/ci_workspaces/MODULE.bazel.tmpl b/toolchains/ci_workspaces/MODULE.bazel.tmpl index 142a021609..66b5d448ca 100644 --- a/toolchains/ci_workspaces/MODULE.bazel.tmpl +++ b/toolchains/ci_workspaces/MODULE.bazel.tmpl @@ -67,20 +67,18 @@ http_archive( http_archive( name = "tensorrt", build_file = "@//third_party/tensorrt/archive:BUILD", - sha256 = "33d3c2f3f4c84dc7991a4337a6fde9ed33f5c8e5c4f03ac2eb6b994a382b03a0", - strip_prefix = "TensorRT-10.6.0.26", + strip_prefix = "TensorRT-10.7.0.23", urls = [ - "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.6.0/tars/TensorRT-10.6.0.26.Linux.x86_64-gnu.cuda-12.6.tar.gz", + "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.7.0/tars/TensorRT-10.7.0.23.Linux.x86_64-gnu.cuda-12.6.tar.gz", ], ) http_archive( name = "tensorrt_win", build_file = "@//third_party/tensorrt/archive:BUILD", - sha256 = "6c6d92c108a1b3368423e8f69f08d31269830f1e4c9da43b37ba34a176797254", - strip_prefix = "TensorRT-10.6.0.26", + strip_prefix = "TensorRT-10.7.0.23", urls = [ - "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.6.0/zip/TensorRT-10.6.0.26.Windows.win10.cuda-12.6.zip", + "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.7.0/zip/TensorRT-10.7.0.23.Windows.win10.cuda-12.6.zip", ], ) diff --git a/toolchains/ci_workspaces/MODULE_tensorrt.bazel.tmpl b/toolchains/ci_workspaces/MODULE_tensorrt.bazel.tmpl index aff07a0383..0a9bb4b9ec 100644 --- a/toolchains/ci_workspaces/MODULE_tensorrt.bazel.tmpl +++ b/toolchains/ci_workspaces/MODULE_tensorrt.bazel.tmpl @@ -67,7 +67,6 @@ http_archive( http_archive( name = "tensorrt", build_file = "@//third_party/tensorrt/archive:BUILD", - sha256 = "${TENSORRT_SHA256}", strip_prefix = "${TENSORRT_STRIP_PREFIX}", urls = [ "${TENSORRT_URLS}", @@ -77,7 +76,6 @@ http_archive( http_archive( name = "tensorrt_win", build_file = "@//third_party/tensorrt/archive:BUILD", - sha256 = "${TENSORRT_SHA256}", strip_prefix = "${TENSORRT_STRIP_PREFIX}", urls = [ "${TENSORRT_URLS}", diff --git a/toolchains/legacy/WORKSPACE.win.release.tmpl b/toolchains/legacy/WORKSPACE.win.release.tmpl index 58fce5cf54..5570516c6c 100644 --- a/toolchains/legacy/WORKSPACE.win.release.tmpl +++ b/toolchains/legacy/WORKSPACE.win.release.tmpl @@ -63,7 +63,7 @@ http_archive( new_local_repository( name = "tensorrt_win", - path = "C:/TensorRT-10.3.0.26", + path = "C:/TensorRT-10.7.0.23", build_file = "@//third_party/tensorrt/local:BUILD" ) diff --git a/toolchains/legacy/WORKSPACE.x86_64.release.rhel.tmpl b/toolchains/legacy/WORKSPACE.x86_64.release.rhel.tmpl index 97b3a8c566..fcc5e33f44 100644 --- a/toolchains/legacy/WORKSPACE.x86_64.release.rhel.tmpl +++ b/toolchains/legacy/WORKSPACE.x86_64.release.rhel.tmpl @@ -71,10 +71,9 @@ http_archive( http_archive( name = "tensorrt", build_file = "@//third_party/tensorrt/archive:BUILD", - sha256 = "adff1cd5abe5d87013806172351e58fd024e5bf0fc61d49ef4b84cd38ed99081", - strip_prefix = "TensorRT-10.3.0.26", + strip_prefix = "TensorRT-10.7.0.23", urls = [ - "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.3.0/tars/TensorRT-10.3.0.26.Linux.x86_64-gnu.cuda-12.5.tar.gz", + "https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.7.0/tars/TensorRT-10.7.0.23.Linux.x86_64-gnu.cuda-12.6.tar.gz", ], ) diff --git a/uv.lock b/uv.lock index c4d4f776f4..cf0a651463 100644 --- a/uv.lock +++ b/uv.lock @@ -1,9 +1,11 @@ version = 1 requires-python = ">=3.9" resolution-markers = [ - "python_full_version < '3.12' and sys_platform == 'linux'", + "python_full_version < '3.11' and sys_platform == 'linux'", + "python_full_version == '3.11.*' and sys_platform == 'linux'", "python_full_version >= '3.12' and sys_platform == 'linux'", - "python_full_version < '3.12' and sys_platform == 'windows'", + "python_full_version < '3.11' and sys_platform == 'windows'", + "python_full_version == '3.11.*' and sys_platform == 'windows'", "python_full_version >= '3.12' and sys_platform == 'windows'", ] supported-markers = [ @@ -13,7 +15,7 @@ supported-markers = [ [[package]] name = "accelerate" -version = "1.2.0" +version = "1.2.1" source = { registry = "https://pypi.org/simple" } dependencies = [ { name = "huggingface-hub", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, @@ -24,9 +26,107 @@ dependencies = [ { name = "safetensors", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, { name = "torch", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, ] -sdist = { url = "https://files.pythonhosted.org/packages/d5/8e/a9827d6df00381f4815ddf84d79bb089783a569e29665e4cf03e808ddebc/accelerate-1.2.0.tar.gz", hash = "sha256:4187a9dea21101255ada3a1974532585feeb6ca8b681693dba362d6d25e85ef2", size = 341495 } +sdist = { url = "https://files.pythonhosted.org/packages/42/09/7947691b7d44bfc739da4a44cc47d6a6d75e6fe9adf047c5234d7cb6be64/accelerate-1.2.1.tar.gz", hash = "sha256:03e161fc69d495daf2b9b5c8d5b43d06e2145520c04727b5bda56d49f1a43ab5", size = 341652 } wheels = [ - { url = "https://files.pythonhosted.org/packages/7b/92/e3f810d0910a71979fe7341803e187cbf85e2b50146c8aa3796eb7bc51b4/accelerate-1.2.0-py3-none-any.whl", hash = "sha256:815708367eb83cd682e120da634b2cf318729bff95f30e2f046c086aae20201e", size = 336311 }, + { url = "https://files.pythonhosted.org/packages/c2/60/a585c806d6c0ec5f8149d44eb202714792802f484e6e2b1bf96b23bd2b00/accelerate-1.2.1-py3-none-any.whl", hash = "sha256:be1cbb958cf837e7cdfbde46b812964b1b8ae94c9c7d94d921540beafcee8ddf", size = 336355 }, +] + +[[package]] +name = "aenum" +version = "3.1.15" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/d0/f8/33e75863394f42e429bb553e05fda7c59763f0fd6848de847a25b3fbccf6/aenum-3.1.15.tar.gz", hash = "sha256:8cbd76cd18c4f870ff39b24284d3ea028fbe8731a58df3aa581e434c575b9559", size = 134730 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/d0/fa/ca0c66b388624ba9dbbf35aab3a9f326bfdf5e56a7237fe8f1b600da6864/aenum-3.1.15-py3-none-any.whl", hash = "sha256:e0dfaeea4c2bd362144b87377e2c61d91958c5ed0b4daf89cb6f45ae23af6288", size = 137633 }, +] + +[[package]] +name = "aiohappyeyeballs" +version = "2.4.4" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/7f/55/e4373e888fdacb15563ef6fa9fa8c8252476ea071e96fb46defac9f18bf2/aiohappyeyeballs-2.4.4.tar.gz", hash = "sha256:5fdd7d87889c63183afc18ce9271f9b0a7d32c2303e394468dd45d514a757745", size = 21977 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/b9/74/fbb6559de3607b3300b9be3cc64e97548d55678e44623db17820dbd20002/aiohappyeyeballs-2.4.4-py3-none-any.whl", hash = "sha256:a980909d50efcd44795c4afeca523296716d50cd756ddca6af8c65b996e27de8", size = 14756 }, +] + +[[package]] +name = "aiohttp" +version = "3.11.11" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "aiohappyeyeballs", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "aiosignal", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "async-timeout", marker = "(python_full_version < '3.11' and sys_platform == 'linux') or (python_full_version < '3.11' and sys_platform == 'windows')" }, + { name = "attrs", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "frozenlist", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "multidict", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "propcache", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "yarl", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/fe/ed/f26db39d29cd3cb2f5a3374304c713fe5ab5a0e4c8ee25a0c45cc6adf844/aiohttp-3.11.11.tar.gz", hash = "sha256:bb49c7f1e6ebf3821a42d81d494f538107610c3a705987f53068546b0e90303e", size = 7669618 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/d8/75/0cdf014b816867d86c0bc26f3d3e3f194198dbf33037890beed629cd4f8f/aiohttp-3.11.11-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:cb23d8bb86282b342481cad4370ea0853a39e4a32a0042bb52ca6bdde132df43", size = 1584635 }, + { url = "https://files.pythonhosted.org/packages/df/2f/95b8f4e4dfeb57c1d9ad9fa911ede35a0249d75aa339edd2c2270dc539da/aiohttp-3.11.11-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:f047569d655f81cb70ea5be942ee5d4421b6219c3f05d131f64088c73bb0917f", size = 1632363 }, + { url = "https://files.pythonhosted.org/packages/39/cb/70cf69ea7c50f5b0021a84f4c59c3622b2b3b81695f48a2f0e42ef7eba6e/aiohttp-3.11.11-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:dd7659baae9ccf94ae5fe8bfaa2c7bc2e94d24611528395ce88d009107e00c6d", size = 1668315 }, + { url = "https://files.pythonhosted.org/packages/2f/cc/3a3fc7a290eabc59839a7e15289cd48f33dd9337d06e301064e1e7fb26c5/aiohttp-3.11.11-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:af01e42ad87ae24932138f154105e88da13ce7d202a6de93fafdafb2883a00ef", size = 1589546 }, + { url = "https://files.pythonhosted.org/packages/15/b4/0f7b0ed41ac6000e283e7332f0f608d734b675a8509763ca78e93714cfb0/aiohttp-3.11.11-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5854be2f3e5a729800bac57a8d76af464e160f19676ab6aea74bde18ad19d438", size = 1544581 }, + { url = "https://files.pythonhosted.org/packages/58/b9/4d06470fd85c687b6b0e31935ef73dde6e31767c9576d617309a2206556f/aiohttp-3.11.11-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:6526e5fb4e14f4bbf30411216780c9967c20c5a55f2f51d3abd6de68320cc2f3", size = 1529256 }, + { url = "https://files.pythonhosted.org/packages/61/a2/6958b1b880fc017fd35f5dfb2c26a9a50c755b75fd9ae001dc2236a4fb79/aiohttp-3.11.11-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:85992ee30a31835fc482468637b3e5bd085fa8fe9392ba0bdcbdc1ef5e9e3c55", size = 1536592 }, + { url = "https://files.pythonhosted.org/packages/0f/dd/b974012a9551fd654f5bb95a6dd3f03d6e6472a17e1a8216dd42e9638d6c/aiohttp-3.11.11-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:88a12ad8ccf325a8a5ed80e6d7c3bdc247d66175afedbe104ee2aaca72960d8e", size = 1607446 }, + { url = "https://files.pythonhosted.org/packages/e0/d3/6c98fd87e638e51f074a3f2061e81fcb92123bcaf1439ac1b4a896446e40/aiohttp-3.11.11-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:0a6d3fbf2232e3a08c41eca81ae4f1dff3d8f1a30bae415ebe0af2d2458b8a33", size = 1628809 }, + { url = "https://files.pythonhosted.org/packages/a8/2e/86e6f85cbca02be042c268c3d93e7f35977a0e127de56e319bdd1569eaa8/aiohttp-3.11.11-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:84a585799c58b795573c7fa9b84c455adf3e1d72f19a2bf498b54a95ae0d194c", size = 1564291 }, + { url = "https://files.pythonhosted.org/packages/be/f9/469588603bd75bf02c8ffb8c8a0d4b217eed446b49d4a767684685aa33fd/aiohttp-3.11.11-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:81b8fe282183e4a3c7a1b72f5ade1094ed1c6345a8f153506d114af5bf8accd9", size = 1685694 }, + { url = "https://files.pythonhosted.org/packages/88/b9/1b7fa43faf6c8616fa94c568dc1309ffee2b6b68b04ac268e5d64b738688/aiohttp-3.11.11-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3af41686ccec6a0f2bdc66686dc0f403c41ac2089f80e2214a0f82d001052c03", size = 1743660 }, + { url = "https://files.pythonhosted.org/packages/2a/8b/0248d19dbb16b67222e75f6aecedd014656225733157e5afaf6a6a07e2e8/aiohttp-3.11.11-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:70d1f9dde0e5dd9e292a6d4d00058737052b01f3532f69c0c65818dac26dc287", size = 1785421 }, + { url = "https://files.pythonhosted.org/packages/c4/11/f478e071815a46ca0a5ae974651ff0c7a35898c55063305a896e58aa1247/aiohttp-3.11.11-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:249cc6912405917344192b9f9ea5cd5b139d49e0d2f5c7f70bdfaf6b4dbf3a2e", size = 1675145 }, + { url = "https://files.pythonhosted.org/packages/26/5d/284d182fecbb5075ae10153ff7374f57314c93a8681666600e3a9e09c505/aiohttp-3.11.11-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:0eb98d90b6690827dcc84c246811feeb4e1eea683c0eac6caed7549be9c84665", size = 1619804 }, + { url = "https://files.pythonhosted.org/packages/1b/78/980064c2ad685c64ce0e8aeeb7ef1e53f43c5b005edcd7d32e60809c4992/aiohttp-3.11.11-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:ec82bf1fda6cecce7f7b915f9196601a1bd1a3079796b76d16ae4cce6d0ef89b", size = 1654007 }, + { url = "https://files.pythonhosted.org/packages/21/8d/9e658d63b1438ad42b96f94da227f2e2c1d5c6001c9e8ffcc0bfb22e9105/aiohttp-3.11.11-cp311-cp311-musllinux_1_2_i686.whl", hash = "sha256:9fd46ce0845cfe28f108888b3ab17abff84ff695e01e73657eec3f96d72eef34", size = 1650022 }, + { url = "https://files.pythonhosted.org/packages/85/fd/a032bf7f2755c2df4f87f9effa34ccc1ef5cea465377dbaeef93bb56bbd6/aiohttp-3.11.11-cp311-cp311-musllinux_1_2_ppc64le.whl", hash = "sha256:bd176afcf8f5d2aed50c3647d4925d0db0579d96f75a31e77cbaf67d8a87742d", size = 1732899 }, + { url = "https://files.pythonhosted.org/packages/c5/0c/c2b85fde167dd440c7ba50af2aac20b5a5666392b174df54c00f888c5a75/aiohttp-3.11.11-cp311-cp311-musllinux_1_2_s390x.whl", hash = "sha256:ec2aa89305006fba9ffb98970db6c8221541be7bee4c1d027421d6f6df7d1ce2", size = 1755142 }, + { url = "https://files.pythonhosted.org/packages/bc/78/91ae1a3b3b3bed8b893c5d69c07023e151b1c95d79544ad04cf68f596c2f/aiohttp-3.11.11-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:92cde43018a2e17d48bb09c79e4d4cb0e236de5063ce897a5e40ac7cb4878773", size = 1692736 }, + { url = "https://files.pythonhosted.org/packages/b7/dd/485061fbfef33165ce7320db36e530cd7116ee1098e9c3774d15a732b3fd/aiohttp-3.11.11-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bd7227b87a355ce1f4bf83bfae4399b1f5bb42e0259cb9405824bd03d2f4336a", size = 1682367 }, + { url = "https://files.pythonhosted.org/packages/e9/d7/9ec5b3ea9ae215c311d88b2093e8da17e67b8856673e4166c994e117ee3e/aiohttp-3.11.11-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d40f9da8cabbf295d3a9dae1295c69975b86d941bc20f0a087f0477fa0a66231", size = 1736989 }, + { url = "https://files.pythonhosted.org/packages/d6/fb/ea94927f7bfe1d86178c9d3e0a8c54f651a0a655214cce930b3c679b8f64/aiohttp-3.11.11-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:ffb3dc385f6bb1568aa974fe65da84723210e5d9707e360e9ecb51f59406cd2e", size = 1793265 }, + { url = "https://files.pythonhosted.org/packages/40/7f/6de218084f9b653026bd7063cd8045123a7ba90c25176465f266976d8c82/aiohttp-3.11.11-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a8f5f7515f3552d899c61202d99dcb17d6e3b0de777900405611cd747cecd1b8", size = 1691841 }, + { url = "https://files.pythonhosted.org/packages/77/e2/992f43d87831cbddb6b09c57ab55499332f60ad6fdbf438ff4419c2925fc/aiohttp-3.11.11-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3499c7ffbfd9c6a3d8d6a2b01c26639da7e43d47c7b4f788016226b1e711caa8", size = 1619317 }, + { url = "https://files.pythonhosted.org/packages/96/74/879b23cdd816db4133325a201287c95bef4ce669acde37f8f1b8669e1755/aiohttp-3.11.11-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:8e2bf8029dbf0810c7bfbc3e594b51c4cc9101fbffb583a3923aea184724203c", size = 1641416 }, + { url = "https://files.pythonhosted.org/packages/30/98/b123f6b15d87c54e58fd7ae3558ff594f898d7f30a90899718f3215ad328/aiohttp-3.11.11-cp312-cp312-musllinux_1_2_i686.whl", hash = "sha256:b6212a60e5c482ef90f2d788835387070a88d52cf6241d3916733c9176d39eab", size = 1646514 }, + { url = "https://files.pythonhosted.org/packages/d7/38/257fda3dc99d6978ab943141d5165ec74fd4b4164baa15e9c66fa21da86b/aiohttp-3.11.11-cp312-cp312-musllinux_1_2_ppc64le.whl", hash = "sha256:d119fafe7b634dbfa25a8c597718e69a930e4847f0b88e172744be24515140da", size = 1702095 }, + { url = "https://files.pythonhosted.org/packages/0c/f4/ddab089053f9fb96654df5505c0a69bde093214b3c3454f6bfdb1845f558/aiohttp-3.11.11-cp312-cp312-musllinux_1_2_s390x.whl", hash = "sha256:6fba278063559acc730abf49845d0e9a9e1ba74f85f0ee6efd5803f08b285853", size = 1734611 }, + { url = "https://files.pythonhosted.org/packages/c3/d6/f30b2bc520c38c8aa4657ed953186e535ae84abe55c08d0f70acd72ff577/aiohttp-3.11.11-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:92fc484e34b733704ad77210c7957679c5c3877bd1e6b6d74b185e9320cc716e", size = 1694576 }, + { url = "https://files.pythonhosted.org/packages/07/db/6d04bc7fd92784900704e16b745484ef45b77bd04e25f58f6febaadf7983/aiohttp-3.11.11-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b63de12e44935d5aca7ed7ed98a255a11e5cb47f83a9fded7a5e41c40277d104", size = 1665178 }, + { url = "https://files.pythonhosted.org/packages/54/5c/e95ade9ae29f375411884d9fd98e50535bf9fe316c9feb0f30cd2ac8f508/aiohttp-3.11.11-cp313-cp313-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:aa54f8ef31d23c506910c21163f22b124facb573bff73930735cf9fe38bf7dff", size = 1717939 }, + { url = "https://files.pythonhosted.org/packages/6f/1c/1e7d5c5daea9e409ed70f7986001b8c9e3a49a50b28404498d30860edab6/aiohttp-3.11.11-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a344d5dc18074e3872777b62f5f7d584ae4344cd6006c17ba12103759d407af3", size = 1775125 }, + { url = "https://files.pythonhosted.org/packages/5d/66/890987e44f7d2f33a130e37e01a164168e6aff06fce15217b6eaf14df4f6/aiohttp-3.11.11-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0b7fb429ab1aafa1f48578eb315ca45bd46e9c37de11fe45c7f5f4138091e2f1", size = 1677176 }, + { url = "https://files.pythonhosted.org/packages/8f/dc/e2ba57d7a52df6cdf1072fd5fa9c6301a68e1cd67415f189805d3eeb031d/aiohttp-3.11.11-cp313-cp313-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c341c7d868750e31961d6d8e60ff040fb9d3d3a46d77fd85e1ab8e76c3e9a5c4", size = 1603192 }, + { url = "https://files.pythonhosted.org/packages/6c/9e/8d08a57de79ca3a358da449405555e668f2c8871a7777ecd2f0e3912c272/aiohttp-3.11.11-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:ed9ee95614a71e87f1a70bc81603f6c6760128b140bc4030abe6abaa988f1c3d", size = 1618296 }, + { url = "https://files.pythonhosted.org/packages/56/51/89822e3ec72db352c32e7fc1c690370e24e231837d9abd056490f3a49886/aiohttp-3.11.11-cp313-cp313-musllinux_1_2_i686.whl", hash = "sha256:de8d38f1c2810fa2a4f1d995a2e9c70bb8737b18da04ac2afbf3971f65781d87", size = 1616524 }, + { url = "https://files.pythonhosted.org/packages/2c/fa/e2e6d9398f462ffaa095e84717c1732916a57f1814502929ed67dd7568ef/aiohttp-3.11.11-cp313-cp313-musllinux_1_2_ppc64le.whl", hash = "sha256:a9b7371665d4f00deb8f32208c7c5e652059b0fda41cf6dbcac6114a041f1cc2", size = 1685471 }, + { url = "https://files.pythonhosted.org/packages/ae/5f/6bb976e619ca28a052e2c0ca7b0251ccd893f93d7c24a96abea38e332bf6/aiohttp-3.11.11-cp313-cp313-musllinux_1_2_s390x.whl", hash = "sha256:620598717fce1b3bd14dd09947ea53e1ad510317c85dda2c9c65b622edc96b12", size = 1715312 }, + { url = "https://files.pythonhosted.org/packages/79/c1/756a7e65aa087c7fac724d6c4c038f2faaa2a42fe56dbc1dd62a33ca7213/aiohttp-3.11.11-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:bf8d9bfee991d8acc72d060d53860f356e07a50f0e0d09a8dfedea1c554dd0d5", size = 1672783 }, + { url = "https://files.pythonhosted.org/packages/0e/17/c8be12436ec19915f67b1ab8240d4105aba0f7e0894a1f0d8939c3e79c70/aiohttp-3.11.11-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1642eceeaa5ab6c9b6dfeaaa626ae314d808188ab23ae196a34c9d97efb68350", size = 1587395 }, + { url = "https://files.pythonhosted.org/packages/43/c0/f4db1ac30ebe855b2fefd6fa98767862d88ac54ab08a6ad07d619146270c/aiohttp-3.11.11-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:2170816e34e10f2fd120f603e951630f8a112e1be3b60963a1f159f5699059a6", size = 1636243 }, + { url = "https://files.pythonhosted.org/packages/ea/a7/9acf20e9a09b0d38b5b55691410500d051a9f4194692cac22b0d0fc92ad9/aiohttp-3.11.11-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:8be8508d110d93061197fd2d6a74f7401f73b6d12f8822bbcd6d74f2b55d71b1", size = 1672323 }, + { url = "https://files.pythonhosted.org/packages/f7/5b/a27e8fe1a3b0e245ca80863eefd83fc00136752d27d2cf1afa0130a76f34/aiohttp-3.11.11-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4eed954b161e6b9b65f6be446ed448ed3921763cc432053ceb606f89d793927e", size = 1589521 }, + { url = "https://files.pythonhosted.org/packages/25/50/8bccd08004e15906791b46f0a908a8e7f5e0c5882b17da96d1933bd34ac0/aiohttp-3.11.11-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d6c9af134da4bc9b3bd3e6a70072509f295d10ee60c697826225b60b9959acdd", size = 1544059 }, + { url = "https://files.pythonhosted.org/packages/84/5a/42250b37b06ee0cb7a03dd1630243b1d739ca3edb5abd8b18f479a539900/aiohttp-3.11.11-cp39-cp39-musllinux_1_2_aarch64.whl", hash = "sha256:44167fc6a763d534a6908bdb2592269b4bf30a03239bcb1654781adf5e49caf1", size = 1530217 }, + { url = "https://files.pythonhosted.org/packages/18/08/eb334da86cd2cdbd0621bb7039255b19ca74ce8b05e8fb61850e2589938c/aiohttp-3.11.11-cp39-cp39-musllinux_1_2_i686.whl", hash = "sha256:479b8c6ebd12aedfe64563b85920525d05d394b85f166b7873c8bde6da612f9c", size = 1536081 }, + { url = "https://files.pythonhosted.org/packages/1a/a9/9d59958084d5bad7e77a44841013bd59768cda94f9f744769461b66038fc/aiohttp-3.11.11-cp39-cp39-musllinux_1_2_ppc64le.whl", hash = "sha256:10b4ff0ad793d98605958089fabfa350e8e62bd5d40aa65cdc69d6785859f94e", size = 1606918 }, + { url = "https://files.pythonhosted.org/packages/4f/e7/27feb1cff17dcddb7a5b703199106196718d622a3aa70f80a386d15361d7/aiohttp-3.11.11-cp39-cp39-musllinux_1_2_s390x.whl", hash = "sha256:b540bd67cfb54e6f0865ceccd9979687210d7ed1a1cc8c01f8e67e2f1e883d28", size = 1629101 }, + { url = "https://files.pythonhosted.org/packages/e8/29/49debcd858b997c655fca274c5247fcfe29bf31a4ddb1ce3f088539b14e4/aiohttp-3.11.11-cp39-cp39-musllinux_1_2_x86_64.whl", hash = "sha256:1dac54e8ce2ed83b1f6b1a54005c87dfed139cf3f777fdc8afc76e7841101226", size = 1567338 }, +] + +[[package]] +name = "aiosignal" +version = "1.3.2" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "frozenlist", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/ba/b5/6d55e80f6d8a08ce22b982eafa278d823b541c925f11ee774b0b9c43473d/aiosignal-1.3.2.tar.gz", hash = "sha256:a8c255c66fafb1e499c9351d0bf32ff2d8a0321595ebac3b93713656d2436f54", size = 19424 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/ec/6a/bc7e17a3e87a2985d3e8f4da4cd0f481060eb78fb08596c42be62c90a4d9/aiosignal-1.3.2-py2.py3-none-any.whl", hash = "sha256:45cde58e409a301715980c2b01d0c28bdde3770d8290b5eb2173759d9acb31a5", size = 7597 }, ] [[package]] @@ -38,6 +138,21 @@ wheels = [ { url = "https://files.pythonhosted.org/packages/78/b6/6307fbef88d9b5ee7421e68d78a9f162e0da4900bc5f5793f6d3d0e34fb8/annotated_types-0.7.0-py3-none-any.whl", hash = "sha256:1f02e8b43a8fbbc3f3e0d4f0f4bfc8131bcb4eebe8849b8e5c773f3a1c582a53", size = 13643 }, ] +[[package]] +name = "anyio" +version = "4.7.0" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "exceptiongroup", marker = "(python_full_version < '3.11' and sys_platform == 'linux') or (python_full_version < '3.11' and sys_platform == 'windows')" }, + { name = "idna", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "sniffio", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "typing-extensions", marker = "(python_full_version < '3.13' and sys_platform == 'linux') or (python_full_version < '3.13' and sys_platform == 'windows')" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/f6/40/318e58f669b1a9e00f5c4453910682e2d9dd594334539c7b7817dabb765f/anyio-4.7.0.tar.gz", hash = "sha256:2f834749c602966b7d456a7567cafcb309f96482b5081d14ac93ccd457f9dd48", size = 177076 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/a0/7a/4daaf3b6c08ad7ceffea4634ec206faeff697526421c20f07628c7372156/anyio-4.7.0-py3-none-any.whl", hash = "sha256:ea60c3723ab42ba6fff7e8ccb0488c898ec538ff4df1f1d5e642c3601d07e352", size = 93052 }, +] + [[package]] name = "asttokens" version = "3.0.0" @@ -47,6 +162,24 @@ wheels = [ { url = "https://files.pythonhosted.org/packages/25/8a/c46dcc25341b5bce5472c718902eb3d38600a903b14fa6aeecef3f21a46f/asttokens-3.0.0-py3-none-any.whl", hash = "sha256:e3078351a059199dd5138cb1c706e6430c05eff2ff136af5eb4790f9d28932e2", size = 26918 }, ] +[[package]] +name = "async-timeout" +version = "5.0.1" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/a5/ae/136395dfbfe00dfc94da3f3e136d0b13f394cba8f4841120e34226265780/async_timeout-5.0.1.tar.gz", hash = "sha256:d9321a7a3d5a6a5e187e824d2fa0793ce379a202935782d555d6e9d2735677d3", size = 9274 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/fe/ba/e2081de779ca30d473f21f5b30e0e737c438205440784c7dfc81efc2b029/async_timeout-5.0.1-py3-none-any.whl", hash = "sha256:39e3809566ff85354557ec2398b55e096c8364bacac9405a7a1fa429e77fe76c", size = 6233 }, +] + +[[package]] +name = "attrs" +version = "24.3.0" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/48/c8/6260f8ccc11f0917360fc0da435c5c9c7504e3db174d5a12a1494887b045/attrs-24.3.0.tar.gz", hash = "sha256:8f5c07333d543103541ba7be0e2ce16eeee8130cb0b3f9238ab904ce1e85baff", size = 805984 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/89/aa/ab0f7891a01eeb2d2e338ae8fecbe57fcebea1a24dbb64d45801bfab481d/attrs-24.3.0-py3-none-any.whl", hash = "sha256:ac96cd038792094f438ad1f6ff80837353805ac950cd2aa0e0625ef19850c308", size = 63397 }, +] + [[package]] name = "black" version = "24.10.0" @@ -70,6 +203,22 @@ wheels = [ { url = "https://files.pythonhosted.org/packages/8d/a7/4b27c50537ebca8bec139b872861f9d2bf501c5ec51fcf897cb924d9e264/black-24.10.0-py3-none-any.whl", hash = "sha256:3bb2b7a1f7b685f85b11fed1ef10f8a9148bceb49853e47a294a3dd963c1dd7d", size = 206898 }, ] +[[package]] +name = "build" +version = "1.2.2.post1" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "colorama", marker = "(os_name == 'nt' and sys_platform == 'linux') or (os_name == 'nt' and sys_platform == 'windows')" }, + { name = "importlib-metadata", marker = "(python_full_version < '3.10.2' and sys_platform == 'linux') or (python_full_version < '3.10.2' and sys_platform == 'windows')" }, + { name = "packaging", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "pyproject-hooks", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "tomli", marker = "(python_full_version < '3.11' and sys_platform == 'linux') or (python_full_version < '3.11' and sys_platform == 'windows')" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/7d/46/aeab111f8e06793e4f0e421fcad593d547fb8313b50990f31681ee2fb1ad/build-1.2.2.post1.tar.gz", hash = "sha256:b36993e92ca9375a219c99e606a122ff365a760a2d4bba0caa09bd5278b608b7", size = 46701 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/84/c2/80633736cd183ee4a62107413def345f7e6e3c01563dbca1417363cf957e/build-1.2.2.post1-py3-none-any.whl", hash = "sha256:1d61c0887fa860c01971625baae8bdd338e517b836a2f70dd1f7aa3a6b2fc5b5", size = 22950 }, +] + [[package]] name = "certifi" version = "2024.8.30" @@ -78,6 +227,55 @@ wheels = [ { url = "https://download.pytorch.org/whl/nightly/certifi-2024.8.30-py3-none-any.whl" }, ] +[[package]] +name = "cffi" +version = "1.17.1" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "pycparser", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/fc/97/c783634659c2920c3fc70419e3af40972dbaf758daa229a7d6ea6135c90d/cffi-1.17.1.tar.gz", hash = "sha256:1c39c6016c32bc48dd54561950ebd6836e1670f2ae46128f67cf49e789c52824", size = 516621 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/de/cc/4635c320081c78d6ffc2cab0a76025b691a91204f4aa317d568ff9280a2d/cffi-1.17.1-cp310-cp310-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:edae79245293e15384b51f88b00613ba9f7198016a5948b5dddf4917d4d26382", size = 426024 }, + { url = "https://files.pythonhosted.org/packages/b6/7b/3b2b250f3aab91abe5f8a51ada1b717935fdaec53f790ad4100fe2ec64d1/cffi-1.17.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:45398b671ac6d70e67da8e4224a065cec6a93541bb7aebe1b198a61b58c7b702", size = 448188 }, + { url = "https://files.pythonhosted.org/packages/d3/48/1b9283ebbf0ec065148d8de05d647a986c5f22586b18120020452fff8f5d/cffi-1.17.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ad9413ccdeda48c5afdae7e4fa2192157e991ff761e7ab8fdd8926f40b160cc3", size = 455571 }, + { url = "https://files.pythonhosted.org/packages/40/87/3b8452525437b40f39ca7ff70276679772ee7e8b394934ff60e63b7b090c/cffi-1.17.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:5da5719280082ac6bd9aa7becb3938dc9f9cbd57fac7d2871717b1feb0902ab6", size = 436687 }, + { url = "https://files.pythonhosted.org/packages/8d/fb/4da72871d177d63649ac449aec2e8a29efe0274035880c7af59101ca2232/cffi-1.17.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2bb1a08b8008b281856e5971307cc386a8e9c5b625ac297e853d36da6efe9c17", size = 446211 }, + { url = "https://files.pythonhosted.org/packages/ab/a0/62f00bcb411332106c02b663b26f3545a9ef136f80d5df746c05878f8c4b/cffi-1.17.1-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:045d61c734659cc045141be4bae381a41d89b741f795af1dd018bfb532fd0df8", size = 461325 }, + { url = "https://files.pythonhosted.org/packages/36/83/76127035ed2e7e27b0787604d99da630ac3123bfb02d8e80c633f218a11d/cffi-1.17.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:6883e737d7d9e4899a8a695e00ec36bd4e5e4f18fabe0aca0efe0a4b44cdb13e", size = 438784 }, + { url = "https://files.pythonhosted.org/packages/21/81/a6cd025db2f08ac88b901b745c163d884641909641f9b826e8cb87645942/cffi-1.17.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:6b8b4a92e1c65048ff98cfe1f735ef8f1ceb72e3d5f0c25fdb12087a23da22be", size = 461564 }, + { url = "https://files.pythonhosted.org/packages/94/dd/a3f0118e688d1b1a57553da23b16bdade96d2f9bcda4d32e7d2838047ff7/cffi-1.17.1-cp311-cp311-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f75c7ab1f9e4aca5414ed4d8e5c0e303a34f4421f8a0d47a4d019ceff0ab6af4", size = 445259 }, + { url = "https://files.pythonhosted.org/packages/2e/ea/70ce63780f096e16ce8588efe039d3c4f91deb1dc01e9c73a287939c79a6/cffi-1.17.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a1ed2dd2972641495a3ec98445e09766f077aee98a1c896dcb4ad0d303628e41", size = 469200 }, + { url = "https://files.pythonhosted.org/packages/1c/a0/a4fa9f4f781bda074c3ddd57a572b060fa0df7655d2a4247bbe277200146/cffi-1.17.1-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:46bf43160c1a35f7ec506d254e5c890f3c03648a4dbac12d624e4490a7046cd1", size = 477235 }, + { url = "https://files.pythonhosted.org/packages/62/12/ce8710b5b8affbcdd5c6e367217c242524ad17a02fe5beec3ee339f69f85/cffi-1.17.1-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a24ed04c8ffd54b0729c07cee15a81d964e6fee0e3d4d342a27b020d22959dc6", size = 459721 }, + { url = "https://files.pythonhosted.org/packages/ff/6b/d45873c5e0242196f042d555526f92aa9e0c32355a1be1ff8c27f077fd37/cffi-1.17.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:610faea79c43e44c71e1ec53a554553fa22321b65fae24889706c0a84d4ad86d", size = 467242 }, + { url = "https://files.pythonhosted.org/packages/1a/52/d9a0e523a572fbccf2955f5abe883cfa8bcc570d7faeee06336fbd50c9fc/cffi-1.17.1-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:a9b15d491f3ad5d692e11f6b71f7857e7835eb677955c00cc0aefcd0669adaf6", size = 477999 }, + { url = "https://files.pythonhosted.org/packages/44/74/f2a2460684a1a2d00ca799ad880d54652841a780c4c97b87754f660c7603/cffi-1.17.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:de2ea4b5833625383e464549fec1bc395c1bdeeb5f25c4a3a82b5a8c756ec22f", size = 454242 }, + { url = "https://files.pythonhosted.org/packages/f8/4a/34599cac7dfcd888ff54e801afe06a19c17787dfd94495ab0c8d35fe99fb/cffi-1.17.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:fc48c783f9c87e60831201f2cce7f3b2e4846bf4d8728eabe54d60700b318a0b", size = 478604 }, + { url = "https://files.pythonhosted.org/packages/cc/b6/db007700f67d151abadf508cbfd6a1884f57eab90b1bb985c4c8c02b0f28/cffi-1.17.1-cp312-cp312-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1257bdabf294dceb59f5e70c64a3e2f462c30c7ad68092d01bbbfb1c16b1ba36", size = 454803 }, + { url = "https://files.pythonhosted.org/packages/1a/df/f8d151540d8c200eb1c6fba8cd0dfd40904f1b0682ea705c36e6c2e97ab3/cffi-1.17.1-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:da95af8214998d77a98cc14e3a3bd00aa191526343078b530ceb0bd710fb48a5", size = 478850 }, + { url = "https://files.pythonhosted.org/packages/28/c0/b31116332a547fd2677ae5b78a2ef662dfc8023d67f41b2a83f7c2aa78b1/cffi-1.17.1-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d63afe322132c194cf832bfec0dc69a99fb9bb6bbd550f161a49e9e855cc78ff", size = 485729 }, + { url = "https://files.pythonhosted.org/packages/91/2b/9a1ddfa5c7f13cab007a2c9cc295b70fbbda7cb10a286aa6810338e60ea1/cffi-1.17.1-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f79fc4fc25f1c8698ff97788206bb3c2598949bfe0fef03d299eb1b5356ada99", size = 471256 }, + { url = "https://files.pythonhosted.org/packages/b2/d5/da47df7004cb17e4955df6a43d14b3b4ae77737dff8bf7f8f333196717bf/cffi-1.17.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b62ce867176a75d03a665bad002af8e6d54644fad99a3c70905c543130e39d93", size = 479424 }, + { url = "https://files.pythonhosted.org/packages/0b/ac/2a28bcf513e93a219c8a4e8e125534f4f6db03e3179ba1c45e949b76212c/cffi-1.17.1-cp312-cp312-musllinux_1_1_aarch64.whl", hash = "sha256:386c8bf53c502fff58903061338ce4f4950cbdcb23e2902d86c0f722b786bbe3", size = 484568 }, + { url = "https://files.pythonhosted.org/packages/d4/38/ca8a4f639065f14ae0f1d9751e70447a261f1a30fa7547a828ae08142465/cffi-1.17.1-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:4ceb10419a9adf4460ea14cfd6bc43d08701f0835e979bf821052f1805850fe8", size = 488736 }, + { url = "https://files.pythonhosted.org/packages/0e/2d/eab2e858a91fdff70533cab61dcff4a1f55ec60425832ddfdc9cd36bc8af/cffi-1.17.1-cp313-cp313-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d01b12eeeb4427d3110de311e1774046ad344f5b1a7403101878976ecd7a10f3", size = 454792 }, + { url = "https://files.pythonhosted.org/packages/75/b2/fbaec7c4455c604e29388d55599b99ebcc250a60050610fadde58932b7ee/cffi-1.17.1-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:706510fe141c86a69c8ddc029c7910003a17353970cff3b904ff0686a5927683", size = 478893 }, + { url = "https://files.pythonhosted.org/packages/4f/b7/6e4a2162178bf1935c336d4da8a9352cccab4d3a5d7914065490f08c0690/cffi-1.17.1-cp313-cp313-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:de55b766c7aa2e2a3092c51e0483d700341182f08e67c63630d5b6f200bb28e5", size = 485810 }, + { url = "https://files.pythonhosted.org/packages/c7/8a/1d0e4a9c26e54746dc08c2c6c037889124d4f59dffd853a659fa545f1b40/cffi-1.17.1-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c59d6e989d07460165cc5ad3c61f9fd8f1b4796eacbd81cee78957842b834af4", size = 471200 }, + { url = "https://files.pythonhosted.org/packages/26/9f/1aab65a6c0db35f43c4d1b4f580e8df53914310afc10ae0397d29d697af4/cffi-1.17.1-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:dd398dbc6773384a17fe0d3e7eeb8d1a21c2200473ee6806bb5e6a8e62bb73dd", size = 479447 }, + { url = "https://files.pythonhosted.org/packages/5f/e4/fb8b3dd8dc0e98edf1135ff067ae070bb32ef9d509d6cb0f538cd6f7483f/cffi-1.17.1-cp313-cp313-musllinux_1_1_aarch64.whl", hash = "sha256:3edc8d958eb099c634dace3c7e16560ae474aa3803a5df240542b305d14e14ed", size = 484358 }, + { url = "https://files.pythonhosted.org/packages/f1/47/d7145bf2dc04684935d57d67dff9d6d795b2ba2796806bb109864be3a151/cffi-1.17.1-cp313-cp313-musllinux_1_1_x86_64.whl", hash = "sha256:72e72408cad3d5419375fc87d289076ee319835bdfa2caad331e377589aebba9", size = 488469 }, + { url = "https://files.pythonhosted.org/packages/ed/65/25a8dc32c53bf5b7b6c2686b42ae2ad58743f7ff644844af7cdb29b49361/cffi-1.17.1-cp39-cp39-manylinux_2_12_i686.manylinux2010_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1d599671f396c4723d016dbddb72fe8e0397082b0a77a4fab8028923bec050e8", size = 424910 }, + { url = "https://files.pythonhosted.org/packages/42/7a/9d086fab7c66bd7c4d0f27c57a1b6b068ced810afc498cc8c49e0088661c/cffi-1.17.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ca74b8dbe6e8e8263c0ffd60277de77dcee6c837a3d0881d8c1ead7268c9e576", size = 447200 }, + { url = "https://files.pythonhosted.org/packages/da/63/1785ced118ce92a993b0ec9e0d0ac8dc3e5dbfbcaa81135be56c69cabbb6/cffi-1.17.1-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:f7f5baafcc48261359e14bcd6d9bff6d4b28d9103847c9e136694cb0501aef87", size = 454565 }, + { url = "https://files.pythonhosted.org/packages/74/06/90b8a44abf3556599cdec107f7290277ae8901a58f75e6fe8f970cd72418/cffi-1.17.1-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:98e3969bcff97cae1b2def8ba499ea3d6f31ddfdb7635374834cf89a1a08ecf0", size = 435635 }, + { url = "https://files.pythonhosted.org/packages/bd/62/a1f468e5708a70b1d86ead5bab5520861d9c7eacce4a885ded9faa7729c3/cffi-1.17.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:cdf5ce3acdfd1661132f2a9c19cac174758dc2352bfe37d98aa7512c6b7178b3", size = 445218 }, + { url = "https://files.pythonhosted.org/packages/5b/95/b34462f3ccb09c2594aa782d90a90b045de4ff1f70148ee79c69d37a0a5a/cffi-1.17.1-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:9755e4345d1ec879e3849e62222a18c7174d65a6a92d5b346b1863912168b595", size = 460486 }, + { url = "https://files.pythonhosted.org/packages/fc/fc/a1e4bebd8d680febd29cf6c8a40067182b64f00c7d105f8f26b5bc54317b/cffi-1.17.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:f1e22e8c4419538cb197e4dd60acc919d7696e5ef98ee4da4e01d3f8cfa4cc5a", size = 437911 }, + { url = "https://files.pythonhosted.org/packages/e6/c3/21cab7a6154b6a5ea330ae80de386e7665254835b9e98ecc1340b3a7de9a/cffi-1.17.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:c03e868a0b3bc35839ba98e74211ed2b05d2119be4e8a0f224fba9384f1fe02e", size = 460632 }, +] + [[package]] name = "cfgv" version = "3.4.0" @@ -130,14 +328,26 @@ wheels = [ [[package]] name = "click" -version = "8.1.7" +version = "8.1.8" source = { registry = "https://pypi.org/simple" } dependencies = [ { name = "colorama", marker = "(platform_system == 'Windows' and sys_platform == 'linux') or (platform_system == 'Windows' and sys_platform == 'windows')" }, ] -sdist = { url = "https://files.pythonhosted.org/packages/96/d3/f04c7bfcf5c1862a2a5b845c6b2b360488cf47af55dfa79c98f6a6bf98b5/click-8.1.7.tar.gz", hash = "sha256:ca9853ad459e787e2192211578cc907e7594e294c7ccc834310722b41b9ca6de", size = 336121 } +sdist = { url = "https://files.pythonhosted.org/packages/b9/2e/0090cbf739cee7d23781ad4b89a9894a41538e4fcf4c31dcdd705b78eb8b/click-8.1.8.tar.gz", hash = "sha256:ed53c9d8990d83c2a27deae68e4ee337473f6330c040a31d4225c9574d16096a", size = 226593 } wheels = [ - { url = "https://files.pythonhosted.org/packages/00/2e/d53fa4befbf2cfa713304affc7ca780ce4fc1fd8710527771b58311a3229/click-8.1.7-py3-none-any.whl", hash = "sha256:ae74fb96c20a0277a1d615f1e4d73c8414f5a98db8b799a7931d1582f3390c28", size = 97941 }, + { url = "https://files.pythonhosted.org/packages/7e/d4/7ebdbd03970677812aac39c869717059dbb71a4cfc033ca6e5221787892c/click-8.1.8-py3-none-any.whl", hash = "sha256:63c132bbbed01578a06712a2d1f497bb62d9c1c0d329b7903a866228027263b2", size = 98188 }, +] + +[[package]] +name = "click-option-group" +version = "0.5.6" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "click", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/e7/b8/91054601a2e05fd9060cb1baf56be5b24145817b059e078669e1099529c7/click-option-group-0.5.6.tar.gz", hash = "sha256:97d06703873518cc5038509443742b25069a3c7562d1ea72ff08bfadde1ce777", size = 16517 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/af/75/81ea958bc0f7e410257cb2a42531b93a7695a31930cde87192c010a52c50/click_option_group-0.5.6-py3-none-any.whl", hash = "sha256:38a26d963ee3ad93332ddf782f9259c5bdfe405e73408d943ef5e7d0c3767ec7", size = 12467 }, ] [[package]] @@ -157,6 +367,27 @@ wheels = [ { url = "https://download.pytorch.org/whl/nightly/colorama-0.4.6-py2.py3-none-any.whl" }, ] +[[package]] +name = "colored" +version = "2.2.4" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/2f/98/4d4546307039955eec131cf9538732fb7a28d2db2090cd1d4e4a135829e1/colored-2.2.4.tar.gz", hash = "sha256:595e1dd7f3b472ea5f12af21d2fec8a2ea2cf8f9d93e67180197330b26df9b61", size = 13202 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/75/d1/548f697f88872321525e294f8863efbdd1c313964b7f94e49ab0dc4f2895/colored-2.2.4-py3-none-any.whl", hash = "sha256:a7069673bd90a35f46cb748d012c17284a0668d2f1c06bc7a51822a2d5ad2112", size = 16109 }, +] + +[[package]] +name = "coloredlogs" +version = "15.0.1" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "humanfriendly", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/cc/c7/eed8f27100517e8c0e6b923d5f0845d0cb99763da6fdee00478f91db7325/coloredlogs-15.0.1.tar.gz", hash = "sha256:7c991aa71a4577af2f82600d8f8f3a89f936baeaf9b50a9c197da014e5bf16b0", size = 278520 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/a7/06/3d6badcf13db419e25b07041d9c7b4a2c331d3f4e7134445ec5df57714cd/coloredlogs-15.0.1-py2.py3-none-any.whl", hash = "sha256:612ee75c546f53e92e70049c9dbfcc18c935a2b9a53b66085ce9ef6a6e5c0934", size = 46018 }, +] + [[package]] name = "comm" version = "0.2.2" @@ -169,6 +400,45 @@ wheels = [ { url = "https://files.pythonhosted.org/packages/e6/75/49e5bfe642f71f272236b5b2d2691cf915a7283cc0ceda56357b61daa538/comm-0.2.2-py3-none-any.whl", hash = "sha256:e6fb86cb70ff661ee8c9c14e7d36d6de3b4066f1441be4063df9c5009f0a64d3", size = 7180 }, ] +[[package]] +name = "cuda-python" +version = "12.6.2.post1" +source = { registry = "https://pypi.org/simple" } +wheels = [ + { url = "https://files.pythonhosted.org/packages/5d/b4/5905cc801fd02a150204c97e18bbdabe4d167b84f7453d241c080128a576/cuda_python-12.6.2.post1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:be1f268aae08d509e4af2d8b8465b74351de39d8f439a5a98caf9b276e027d9b", size = 24497758 }, + { url = "https://files.pythonhosted.org/packages/7d/8e/74c493c3bd3855509bdca3001e84d0b4af499cbc83a71f3b4d42ba827781/cuda_python-12.6.2.post1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ee91c92f34fc140b8e241a2681747cfb4442fa3a9dc817376d3090f6c73a0c0f", size = 24910886 }, + { url = "https://files.pythonhosted.org/packages/ec/a1/9a00f7ace9ca34c8b54a8a21c5406a602daadebdb93fd0bddf886c0ec8e8/cuda_python-12.6.2.post1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:20271f7979495435e2192758ca52a227dc70e04af1453442a44d3149b94c4630", size = 25349718 }, + { url = "https://files.pythonhosted.org/packages/4e/f4/9badcb1f59263365a2ce49b09b4a0bfa95c1c102a367c5601bcfe118cd96/cuda_python-12.6.2.post1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d47ba3464fd890025a6d1929dac5e71f2d8c16d0abfe461123cf8be1f975d3a0", size = 25742930 }, + { url = "https://files.pythonhosted.org/packages/d5/c4/4cc082de59d4a465cd29125853b916d5b229a6961cc173a0f22ee699621b/cuda_python-12.6.2.post1-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ea1e5b26cdfa424c26f098f6ec17340608b73dddc5fd2059f5b31897eabd9916", size = 24612106 }, + { url = "https://files.pythonhosted.org/packages/95/70/7fc0e460a4d2f24b79854cbe7cd474a5457f4b849dd9cd794fafeb4fb85a/cuda_python-12.6.2.post1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1ccda3b9400dd6568b3db118aae1c783ef26f2457d6635197999fa42b46b5187", size = 24750651 }, + { url = "https://files.pythonhosted.org/packages/4f/6e/0583be6c547b81e8a3790b8ea845776c2655641b5cae247aacb0bf5ebf51/cuda_python-12.6.2.post1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3ea25bf01b08563b2d9ff207db1f6c9beb21cd79528fc3f818cf1771c6a05306", size = 24473514 }, + { url = "https://files.pythonhosted.org/packages/ab/ab/51c55ca37cc4622e6bf6953a0fb90ea1f871ab5e090fc01082a2bb402443/cuda_python-12.6.2.post1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6dfb5bdafdc26b47cd345861e6374843bede99fe81804377404c35c19e59e21d", size = 24912694 }, +] + +[[package]] +name = "datasets" +version = "2.14.4" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "aiohttp", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "dill", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "fsspec", extra = ["http"], marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "huggingface-hub", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "multiprocess", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "numpy", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "packaging", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "pandas", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "pyarrow", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "pyyaml", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "requests", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "tqdm", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "xxhash", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/1d/69/8cc725b5d38968fd118e4ce56a483b16e75b7793854c1a392ec4a34eeb31/datasets-2.14.4.tar.gz", hash = "sha256:ef29c2b5841de488cd343cfc26ab979bff77efa4d2285af51f1ad7db5c46a83b", size = 2178719 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/66/f8/38298237d18d4b6a8ee5dfe390e97bed5adb8e01ec6f9680c0ddf3066728/datasets-2.14.4-py3-none-any.whl", hash = "sha256:29336bd316a7d827ccd4da2236596279b20ca2ac78f64c04c9483da7cbc2459b", size = 519335 }, +] + [[package]] name = "decorator" version = "5.1.1" @@ -180,7 +450,7 @@ wheels = [ [[package]] name = "diffusers" -version = "0.31.0" +version = "0.32.0" source = { registry = "https://pypi.org/simple" } dependencies = [ { name = "filelock", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, @@ -192,9 +462,18 @@ dependencies = [ { name = "requests", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, { name = "safetensors", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, ] -sdist = { url = "https://files.pythonhosted.org/packages/78/5d/156acb741303abbee214926804c5f0d09eacd35d05ad942577e996acdac3/diffusers-0.31.0.tar.gz", hash = "sha256:b1d01a73e45d43a0630c299173915dddd69fc50f2ae8f2ab5de4fd245eaed72f", size = 2308202 } +sdist = { url = "https://files.pythonhosted.org/packages/12/fa/48b5be99873a1e5916663c0baab408cb5b74b0a060854e5ff06b54b7630c/diffusers-0.32.0.tar.gz", hash = "sha256:a99ebc2e645b9417e06646bc4c82e26099b4dd9d01b4b717153b061bd62b1032", size = 2617524 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/7b/d8/0fe86ea4c22bc555b63b72148bde055cf32cd0b839bdc4aee0eece7f4a92/diffusers-0.32.0-py3-none-any.whl", hash = "sha256:81d878e9fb6515c3155d3698168102d637b86ccad8bba44c418fefb6b5ed5428", size = 3224169 }, +] + +[[package]] +name = "dill" +version = "0.3.7" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/c4/31/54dd222e02311c2dbc9e680d37cbd50f4494ce1ee9b04c69980e4ec26f38/dill-0.3.7.tar.gz", hash = "sha256:cc1c8b182eb3013e24bd475ff2e9295af86c1a38eb1aff128dac8962a9ce3c03", size = 183355 } wheels = [ - { url = "https://files.pythonhosted.org/packages/07/7f/53828ff41b86c3274099327627ff5cb7581a2cc9db3bcb39e5212c4140b7/diffusers-0.31.0-py3-none-any.whl", hash = "sha256:cbc498ae63f4abfc7c3a07649cdcbee229ef2f9a9a1f0d19c9bbaf22f8d30c1f", size = 2873496 }, + { url = "https://files.pythonhosted.org/packages/f5/3a/74a29b11cf2cdfcd6ba89c0cecd70b37cd1ba7b77978ce611eb7a146a832/dill-0.3.7-py3-none-any.whl", hash = "sha256:76b122c08ef4ce2eedcd4d1abd8e641114bfc6c2867f49f3c41facf65bf19f5e", size = 115254 }, ] [[package]] @@ -206,6 +485,37 @@ wheels = [ { url = "https://files.pythonhosted.org/packages/91/a1/cf2472db20f7ce4a6be1253a81cfdf85ad9c7885ffbed7047fb72c24cf87/distlib-0.3.9-py2.py3-none-any.whl", hash = "sha256:47f8c22fd27c27e25a65601af709b38e4f0a45ea4fc2e710f65755fa8caaaf87", size = 468973 }, ] +[[package]] +name = "distro" +version = "1.9.0" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/fc/f8/98eea607f65de6527f8a2e8885fc8015d3e6f5775df186e443e0964a11c3/distro-1.9.0.tar.gz", hash = "sha256:2fa77c6fd8940f116ee1d6b94a2f90b13b5ea8d019b98bc8bafdcabcdd9bdbed", size = 60722 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/12/b3/231ffd4ab1fc9d679809f356cebee130ac7daa00d6d6f3206dd4fd137e9e/distro-1.9.0-py3-none-any.whl", hash = "sha256:7bffd925d65168f85027d8da9af6bddab658135b840670a223589bc0c8ef02b2", size = 20277 }, +] + +[[package]] +name = "evaluate" +version = "0.4.3" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "datasets", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "dill", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "fsspec", extra = ["http"], marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "huggingface-hub", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "multiprocess", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "numpy", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "packaging", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "pandas", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "requests", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "tqdm", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "xxhash", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/5a/a0/10a56e0939ece94c54276e81459cb4101f46f0e9a6f54fc31a35f64e8854/evaluate-0.4.3.tar.gz", hash = "sha256:3a5700cf83aabee9549264e1e5666f116367c61dbd4d38352015e859a5e2098d", size = 65679 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/a2/e7/cbca9e2d2590eb9b5aa8f7ebabe1beb1498f9462d2ecede5c9fd9735faaf/evaluate-0.4.3-py3-none-any.whl", hash = "sha256:47d8770bdea76e2c2ed0d40189273027d1a41ccea861bcc7ba12d30ec5d1e517", size = 84010 }, +] + [[package]] name = "exceptiongroup" version = "1.2.2" @@ -242,6 +552,20 @@ wheels = [ { url = "https://files.pythonhosted.org/packages/c7/39/689391845f5dc48df81b0c22248d5f66919b82da12f2bab1424bc3610529/expecttest-0.1.6-py3-none-any.whl", hash = "sha256:7cf2db203c06f9e3173670ca9d09ac00912e535139afac2c7458c1627b1a3ee6", size = 6535 }, ] +[[package]] +name = "fastapi" +version = "0.115.4" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "pydantic", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "starlette", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "typing-extensions", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/a9/db/5781f19bd30745885e0737ff3fdd4e63e7bc691710f9da691128bb0dc73b/fastapi-0.115.4.tar.gz", hash = "sha256:db653475586b091cb8b2fec2ac54a680ac6a158e07406e1abae31679e8826349", size = 300737 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/99/f6/af0d1f58f86002be0cf1e2665cdd6f7a4a71cdc8a7a9438cdc9e3b5375fe/fastapi-0.115.4-py3-none-any.whl", hash = "sha256:0b504a063ffb3cf96a5e27dc1bc32c80ca743a2528574f9cdc77daa2d31b4742", size = 94732 }, +] + [[package]] name = "filelock" version = "3.16.1" @@ -250,6 +574,65 @@ wheels = [ { url = "https://download.pytorch.org/whl/nightly/filelock-3.16.1-py3-none-any.whl" }, ] +[[package]] +name = "frozenlist" +version = "1.5.0" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/8f/ed/0f4cec13a93c02c47ec32d81d11c0c1efbadf4a471e3f3ce7cad366cbbd3/frozenlist-1.5.0.tar.gz", hash = "sha256:81d5af29e61b9c8348e876d442253723928dce6433e0e76cd925cd83f1b4b817", size = 39930 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/17/62/594a6829ac5679c25755362a9dc93486a8a45241394564309641425d3ff6/frozenlist-1.5.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e79225373c317ff1e35f210dd5f1344ff31066ba8067c307ab60254cd3a78ad5", size = 240946 }, + { url = "https://files.pythonhosted.org/packages/7e/75/6c8419d8f92c80dd0ee3f63bdde2702ce6398b0ac8410ff459f9b6f2f9cb/frozenlist-1.5.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:9272fa73ca71266702c4c3e2d4a28553ea03418e591e377a03b8e3659d94fa76", size = 264608 }, + { url = "https://files.pythonhosted.org/packages/88/3e/82a6f0b84bc6fb7e0be240e52863c6d4ab6098cd62e4f5b972cd31e002e8/frozenlist-1.5.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:498524025a5b8ba81695761d78c8dd7382ac0b052f34e66939c42df860b8ff17", size = 261361 }, + { url = "https://files.pythonhosted.org/packages/fd/85/14e5f9ccac1b64ff2f10c927b3ffdf88772aea875882406f9ba0cec8ad84/frozenlist-1.5.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:92b5278ed9d50fe610185ecd23c55d8b307d75ca18e94c0e7de328089ac5dcba", size = 231649 }, + { url = "https://files.pythonhosted.org/packages/ee/59/928322800306f6529d1852323014ee9008551e9bb027cc38d276cbc0b0e7/frozenlist-1.5.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7f3c8c1dacd037df16e85227bac13cca58c30da836c6f936ba1df0c05d046d8d", size = 241853 }, + { url = "https://files.pythonhosted.org/packages/7d/bd/e01fa4f146a6f6c18c5d34cab8abdc4013774a26c4ff851128cd1bd3008e/frozenlist-1.5.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:f2ac49a9bedb996086057b75bf93538240538c6d9b38e57c82d51f75a73409d2", size = 243652 }, + { url = "https://files.pythonhosted.org/packages/a5/bd/e4771fd18a8ec6757033f0fa903e447aecc3fbba54e3630397b61596acf0/frozenlist-1.5.0-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:e66cc454f97053b79c2ab09c17fbe3c825ea6b4de20baf1be28919460dd7877f", size = 241734 }, + { url = "https://files.pythonhosted.org/packages/21/13/c83821fa5544af4f60c5d3a65d054af3213c26b14d3f5f48e43e5fb48556/frozenlist-1.5.0-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:5a3ba5f9a0dfed20337d3e966dc359784c9f96503674c2faf015f7fe8e96798c", size = 260959 }, + { url = "https://files.pythonhosted.org/packages/71/f3/1f91c9a9bf7ed0e8edcf52698d23f3c211d8d00291a53c9f115ceb977ab1/frozenlist-1.5.0-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:6321899477db90bdeb9299ac3627a6a53c7399c8cd58d25da094007402b039ab", size = 262706 }, + { url = "https://files.pythonhosted.org/packages/4c/22/4a256fdf5d9bcb3ae32622c796ee5ff9451b3a13a68cfe3f68e2c95588ce/frozenlist-1.5.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:76e4753701248476e6286f2ef492af900ea67d9706a0155335a40ea21bf3b2f5", size = 250401 }, + { url = "https://files.pythonhosted.org/packages/98/a8/d0ac0b9276e1404f58fec3ab6e90a4f76b778a49373ccaf6a563f100dfbc/frozenlist-1.5.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0996c66760924da6e88922756d99b47512a71cfd45215f3570bf1e0b694c206a", size = 276357 }, + { url = "https://files.pythonhosted.org/packages/ad/c9/c7761084fa822f07dac38ac29f841d4587570dd211e2262544aa0b791d21/frozenlist-1.5.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:a2fe128eb4edeabe11896cb6af88fca5346059f6c8d807e3b910069f39157869", size = 287516 }, + { url = "https://files.pythonhosted.org/packages/a1/ff/cd7479e703c39df7bdab431798cef89dc75010d8aa0ca2514c5b9321db27/frozenlist-1.5.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:1a8ea951bbb6cacd492e3948b8da8c502a3f814f5d20935aae74b5df2b19cf3d", size = 283131 }, + { url = "https://files.pythonhosted.org/packages/59/a0/370941beb47d237eca4fbf27e4e91389fd68699e6f4b0ebcc95da463835b/frozenlist-1.5.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:de537c11e4aa01d37db0d403b57bd6f0546e71a82347a97c6a9f0dcc532b3a45", size = 261320 }, + { url = "https://files.pythonhosted.org/packages/b8/5f/c10123e8d64867bc9b4f2f510a32042a306ff5fcd7e2e09e5ae5100ee333/frozenlist-1.5.0-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9c2623347b933fcb9095841f1cc5d4ff0b278addd743e0e966cb3d460278840d", size = 274877 }, + { url = "https://files.pythonhosted.org/packages/fa/79/38c505601ae29d4348f21706c5d89755ceded02a745016ba2f58bd5f1ea6/frozenlist-1.5.0-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:cee6798eaf8b1416ef6909b06f7dc04b60755206bddc599f52232606e18179d3", size = 269592 }, + { url = "https://files.pythonhosted.org/packages/19/e2/39f3a53191b8204ba9f0bb574b926b73dd2efba2a2b9d2d730517e8f7622/frozenlist-1.5.0-cp311-cp311-musllinux_1_2_i686.whl", hash = "sha256:f5f9da7f5dbc00a604fe74aa02ae7c98bcede8a3b8b9666f9f86fc13993bc71a", size = 265934 }, + { url = "https://files.pythonhosted.org/packages/d5/c9/3075eb7f7f3a91f1a6b00284af4de0a65a9ae47084930916f5528144c9dd/frozenlist-1.5.0-cp311-cp311-musllinux_1_2_ppc64le.whl", hash = "sha256:90646abbc7a5d5c7c19461d2e3eeb76eb0b204919e6ece342feb6032c9325ae9", size = 283859 }, + { url = "https://files.pythonhosted.org/packages/05/f5/549f44d314c29408b962fa2b0e69a1a67c59379fb143b92a0a065ffd1f0f/frozenlist-1.5.0-cp311-cp311-musllinux_1_2_s390x.whl", hash = "sha256:bdac3c7d9b705d253b2ce370fde941836a5f8b3c5c2b8fd70940a3ea3af7f4f2", size = 287560 }, + { url = "https://files.pythonhosted.org/packages/9d/f8/cb09b3c24a3eac02c4c07a9558e11e9e244fb02bf62c85ac2106d1eb0c0b/frozenlist-1.5.0-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:03d33c2ddbc1816237a67f66336616416e2bbb6beb306e5f890f2eb22b959cdf", size = 277150 }, + { url = "https://files.pythonhosted.org/packages/e3/12/2aad87deb08a4e7ccfb33600871bbe8f0e08cb6d8224371387f3303654d7/frozenlist-1.5.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:feeb64bc9bcc6b45c6311c9e9b99406660a9c05ca8a5b30d14a78555088b0b3a", size = 282647 }, + { url = "https://files.pythonhosted.org/packages/77/f2/07f06b05d8a427ea0060a9cef6e63405ea9e0d761846b95ef3fb3be57111/frozenlist-1.5.0-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:683173d371daad49cffb8309779e886e59c2f369430ad28fe715f66d08d4ab1a", size = 289052 }, + { url = "https://files.pythonhosted.org/packages/bd/9f/8bf45a2f1cd4aa401acd271b077989c9267ae8463e7c8b1eb0d3f561b65e/frozenlist-1.5.0-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:7d57d8f702221405a9d9b40f9da8ac2e4a1a8b5285aac6100f3393675f0a85ee", size = 291719 }, + { url = "https://files.pythonhosted.org/packages/41/d1/1f20fd05a6c42d3868709b7604c9f15538a29e4f734c694c6bcfc3d3b935/frozenlist-1.5.0-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:30c72000fbcc35b129cb09956836c7d7abf78ab5416595e4857d1cae8d6251a6", size = 267433 }, + { url = "https://files.pythonhosted.org/packages/af/f2/64b73a9bb86f5a89fb55450e97cd5c1f84a862d4ff90d9fd1a73ab0f64a5/frozenlist-1.5.0-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:000a77d6034fbad9b6bb880f7ec073027908f1b40254b5d6f26210d2dab1240e", size = 283591 }, + { url = "https://files.pythonhosted.org/packages/29/e2/ffbb1fae55a791fd6c2938dd9ea779509c977435ba3940b9f2e8dc9d5316/frozenlist-1.5.0-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:5d7f5a50342475962eb18b740f3beecc685a15b52c91f7d975257e13e029eca9", size = 273249 }, + { url = "https://files.pythonhosted.org/packages/2e/6e/008136a30798bb63618a114b9321b5971172a5abddff44a100c7edc5ad4f/frozenlist-1.5.0-cp312-cp312-musllinux_1_2_i686.whl", hash = "sha256:87f724d055eb4785d9be84e9ebf0f24e392ddfad00b3fe036e43f489fafc9039", size = 271075 }, + { url = "https://files.pythonhosted.org/packages/ae/f0/4e71e54a026b06724cec9b6c54f0b13a4e9e298cc8db0f82ec70e151f5ce/frozenlist-1.5.0-cp312-cp312-musllinux_1_2_ppc64le.whl", hash = "sha256:6e9080bb2fb195a046e5177f10d9d82b8a204c0736a97a153c2466127de87784", size = 285398 }, + { url = "https://files.pythonhosted.org/packages/4d/36/70ec246851478b1c0b59f11ef8ade9c482ff447c1363c2bd5fad45098b12/frozenlist-1.5.0-cp312-cp312-musllinux_1_2_s390x.whl", hash = "sha256:9b93d7aaa36c966fa42efcaf716e6b3900438632a626fb09c049f6a2f09fc631", size = 294445 }, + { url = "https://files.pythonhosted.org/packages/37/e0/47f87544055b3349b633a03c4d94b405956cf2437f4ab46d0928b74b7526/frozenlist-1.5.0-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:52ef692a4bc60a6dd57f507429636c2af8b6046db8b31b18dac02cbc8f507f7f", size = 280569 }, + { url = "https://files.pythonhosted.org/packages/1f/22/462a3dd093d11df623179d7754a3b3269de3b42de2808cddef50ee0f4f48/frozenlist-1.5.0-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6482a5851f5d72767fbd0e507e80737f9c8646ae7fd303def99bfe813f76cf7f", size = 265636 }, + { url = "https://files.pythonhosted.org/packages/80/cf/e075e407fc2ae7328155a1cd7e22f932773c8073c1fc78016607d19cc3e5/frozenlist-1.5.0-cp313-cp313-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:44c49271a937625619e862baacbd037a7ef86dd1ee215afc298a417ff3270608", size = 270214 }, + { url = "https://files.pythonhosted.org/packages/a1/58/0642d061d5de779f39c50cbb00df49682832923f3d2ebfb0fedf02d05f7f/frozenlist-1.5.0-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:12f78f98c2f1c2429d42e6a485f433722b0061d5c0b0139efa64f396efb5886b", size = 273905 }, + { url = "https://files.pythonhosted.org/packages/ab/66/3fe0f5f8f2add5b4ab7aa4e199f767fd3b55da26e3ca4ce2cc36698e50c4/frozenlist-1.5.0-cp313-cp313-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ce3aa154c452d2467487765e3adc730a8c153af77ad84096bc19ce19a2400840", size = 250542 }, + { url = "https://files.pythonhosted.org/packages/f6/b8/260791bde9198c87a465224e0e2bb62c4e716f5d198fc3a1dacc4895dbd1/frozenlist-1.5.0-cp313-cp313-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9b7dc0c4338e6b8b091e8faf0db3168a37101943e687f373dce00959583f7439", size = 267026 }, + { url = "https://files.pythonhosted.org/packages/2e/a4/3d24f88c527f08f8d44ade24eaee83b2627793fa62fa07cbb7ff7a2f7d42/frozenlist-1.5.0-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:45e0896250900b5aa25180f9aec243e84e92ac84bd4a74d9ad4138ef3f5c97de", size = 257690 }, + { url = "https://files.pythonhosted.org/packages/de/9a/d311d660420b2beeff3459b6626f2ab4fb236d07afbdac034a4371fe696e/frozenlist-1.5.0-cp313-cp313-musllinux_1_2_i686.whl", hash = "sha256:561eb1c9579d495fddb6da8959fd2a1fca2c6d060d4113f5844b433fc02f2641", size = 253893 }, + { url = "https://files.pythonhosted.org/packages/c6/23/e491aadc25b56eabd0f18c53bb19f3cdc6de30b2129ee0bc39cd387cd560/frozenlist-1.5.0-cp313-cp313-musllinux_1_2_ppc64le.whl", hash = "sha256:df6e2f325bfee1f49f81aaac97d2aa757c7646534a06f8f577ce184afe2f0a9e", size = 267006 }, + { url = "https://files.pythonhosted.org/packages/08/c4/ab918ce636a35fb974d13d666dcbe03969592aeca6c3ab3835acff01f79c/frozenlist-1.5.0-cp313-cp313-musllinux_1_2_s390x.whl", hash = "sha256:140228863501b44b809fb39ec56b5d4071f4d0aa6d216c19cbb08b8c5a7eadb9", size = 276157 }, + { url = "https://files.pythonhosted.org/packages/c0/29/3b7a0bbbbe5a34833ba26f686aabfe982924adbdcafdc294a7a129c31688/frozenlist-1.5.0-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:7707a25d6a77f5d27ea7dc7d1fc608aa0a478193823f88511ef5e6b8a48f9d03", size = 264642 }, + { url = "https://files.pythonhosted.org/packages/08/04/e2fddc92135276e07addbc1cf413acffa0c2d848b3e54cacf684e146df49/frozenlist-1.5.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0f253985bb515ecd89629db13cb58d702035ecd8cfbca7d7a7e29a0e6d39af5f", size = 241756 }, + { url = "https://files.pythonhosted.org/packages/c6/52/be5ff200815d8a341aee5b16b6b707355e0ca3652953852238eb92b120c2/frozenlist-1.5.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:04a5c6babd5e8fb7d3c871dc8b321166b80e41b637c31a995ed844a6139942b6", size = 267718 }, + { url = "https://files.pythonhosted.org/packages/88/be/4bd93a58be57a3722fc544c36debdf9dcc6758f761092e894d78f18b8f20/frozenlist-1.5.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a9fe0f1c29ba24ba6ff6abf688cb0b7cf1efab6b6aa6adc55441773c252f7411", size = 263494 }, + { url = "https://files.pythonhosted.org/packages/32/ba/58348b90193caa096ce9e9befea6ae67f38dabfd3aacb47e46137a6250a8/frozenlist-1.5.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:226d72559fa19babe2ccd920273e767c96a49b9d3d38badd7c91a0fdeda8ea08", size = 232838 }, + { url = "https://files.pythonhosted.org/packages/f6/33/9f152105227630246135188901373c4f322cc026565ca6215b063f4c82f4/frozenlist-1.5.0-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:15b731db116ab3aedec558573c1a5eec78822b32292fe4f2f0345b7f697745c2", size = 242912 }, + { url = "https://files.pythonhosted.org/packages/a0/10/3db38fb3ccbafadd80a1b0d6800c987b0e3fe3ef2d117c6ced0246eea17a/frozenlist-1.5.0-cp39-cp39-musllinux_1_2_aarch64.whl", hash = "sha256:366d8f93e3edfe5a918c874702f78faac300209a4d5bf38352b2c1bdc07a766d", size = 244763 }, + { url = "https://files.pythonhosted.org/packages/e2/cd/1df468fdce2f66a4608dffe44c40cdc35eeaa67ef7fd1d813f99a9a37842/frozenlist-1.5.0-cp39-cp39-musllinux_1_2_i686.whl", hash = "sha256:1b96af8c582b94d381a1c1f51ffaedeb77c821c690ea5f01da3d70a487dd0a9b", size = 242841 }, + { url = "https://files.pythonhosted.org/packages/ee/5f/16097a5ca0bb6b6779c02cc9379c72fe98d56115d4c54d059fb233168fb6/frozenlist-1.5.0-cp39-cp39-musllinux_1_2_ppc64le.whl", hash = "sha256:c03eff4a41bd4e38415cbed054bbaff4a075b093e2394b6915dca34a40d1e38b", size = 263407 }, + { url = "https://files.pythonhosted.org/packages/0f/f7/58cd220ee1c2248ee65a32f5b4b93689e3fe1764d85537eee9fc392543bc/frozenlist-1.5.0-cp39-cp39-musllinux_1_2_s390x.whl", hash = "sha256:50cf5e7ee9b98f22bdecbabf3800ae78ddcc26e4a435515fc72d97903e8488e0", size = 265083 }, + { url = "https://files.pythonhosted.org/packages/62/b8/49768980caabf81ac4a2d156008f7cbd0107e6b36d08a313bb31035d9201/frozenlist-1.5.0-cp39-cp39-musllinux_1_2_x86_64.whl", hash = "sha256:1e76bfbc72353269c44e0bc2cfe171900fbf7f722ad74c9a7b638052afe6a00c", size = 251564 }, + { url = "https://files.pythonhosted.org/packages/c6/c8/a5be5b7550c10858fcf9b0ea054baccab474da77d37f1e828ce043a3a5d4/frozenlist-1.5.0-py3-none-any.whl", hash = "sha256:d994863bba198a4a518b467bb971c56e1db3f180a25c6cf7bb1949c267f748c3", size = 11901 }, +] + [[package]] name = "fsspec" version = "2024.10.0" @@ -258,6 +641,69 @@ wheels = [ { url = "https://download.pytorch.org/whl/nightly/fsspec-2024.10.0-py3-none-any.whl" }, ] +[package.optional-dependencies] +http = [ + { name = "aiohttp", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, +] + +[[package]] +name = "h11" +version = "0.14.0" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/f5/38/3af3d3633a34a3316095b39c8e8fb4853a28a536e55d347bd8d8e9a14b03/h11-0.14.0.tar.gz", hash = "sha256:8f19fbbe99e72420ff35c00b27a34cb9937e902a8b810e2c88300c6f0a3b699d", size = 100418 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/95/04/ff642e65ad6b90db43e668d70ffb6736436c7ce41fcc549f4e9472234127/h11-0.14.0-py3-none-any.whl", hash = "sha256:e3fe4ac4b851c468cc8363d500db52c2ead036020723024a109d37346efaa761", size = 58259 }, +] + +[[package]] +name = "h5py" +version = "3.12.1" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "numpy", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/cc/0c/5c2b0a88158682aeafb10c1c2b735df5bc31f165bfe192f2ee9f2a23b5f1/h5py-3.12.1.tar.gz", hash = "sha256:326d70b53d31baa61f00b8aa5f95c2fcb9621a3ee8365d770c551a13dbbcbfdf", size = 411457 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/1e/d0/4bf67c3937a2437c20844165766ddd1a1817ae6b9544c3743050d8e0f403/h5py-3.12.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3b15d8dbd912c97541312c0e07438864d27dbca857c5ad634de68110c6beb1c2", size = 5168596 }, + { url = "https://files.pythonhosted.org/packages/85/bc/e76f4b2096e0859225f5441d1b7f5e2041fffa19fc2c16756c67078417aa/h5py-3.12.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:59685fe40d8c1fbbee088c88cd4da415a2f8bee5c270337dc5a1c4aa634e3307", size = 5341537 }, + { url = "https://files.pythonhosted.org/packages/b0/62/e2b1f9723ff713e3bd3c16dfeceec7017eadc21ef063d8b7080c0fcdc58a/h5py-3.12.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1473348139b885393125126258ae2d70753ef7e9cec8e7848434f385ae72069e", size = 5273038 }, + { url = "https://files.pythonhosted.org/packages/e1/89/118c3255d6ff2db33b062ec996a762d99ae50c21f54a8a6047ae8eda1b9f/h5py-3.12.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:018a4597f35092ae3fb28ee851fdc756d2b88c96336b8480e124ce1ac6fb9166", size = 5452688 }, + { url = "https://files.pythonhosted.org/packages/af/52/c604adc06280c15a29037d4aa79a24fe54d8d0b51085e81ed24b2fa995f7/h5py-3.12.1-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:050a4f2c9126054515169c49cb900949814987f0c7ae74c341b0c9f9b5056834", size = 5194606 }, + { url = "https://files.pythonhosted.org/packages/fa/63/eeaacff417b393491beebabb8a3dc5342950409eb6d7b39d437289abdbae/h5py-3.12.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5c4b41d1019322a5afc5082864dfd6359f8935ecd37c11ac0029be78c5d112c9", size = 5413256 }, + { url = "https://files.pythonhosted.org/packages/8a/4f/b74332f313bfbe94ba03fff784219b9db385e6139708e55b11490149f90a/h5py-3.12.1-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d3e465aee0ec353949f0f46bf6c6f9790a2006af896cee7c178a8c3e5090aa32", size = 5154390 }, + { url = "https://files.pythonhosted.org/packages/1a/57/93ea9e10a6457ea8d3b867207deb29a527e966a08a84c57ffd954e32152a/h5py-3.12.1-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ba51c0c5e029bb5420a343586ff79d56e7455d496d18a30309616fdbeed1068f", size = 5378244 }, + { url = "https://files.pythonhosted.org/packages/7b/f9/e597b5fef05f161c67a18e8c61bf105209fd242f2612b0ad1aff7ecb0b9c/h5py-3.12.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6fdf6d7936fa824acfa27305fe2d9f39968e539d831c5bae0e0d83ed521ad1ac", size = 5190324 }, + { url = "https://files.pythonhosted.org/packages/8e/1d/631c200e6d5d067035c58028f305cf7f29c494ddfb9b9484a907a367c8bd/h5py-3.12.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:84342bffd1f82d4f036433e7039e241a243531a1d3acd7341b35ae58cdab05bf", size = 5361780 }, +] + +[[package]] +name = "httpcore" +version = "1.0.7" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "certifi", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "h11", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/6a/41/d7d0a89eb493922c37d343b607bc1b5da7f5be7e383740b4753ad8943e90/httpcore-1.0.7.tar.gz", hash = "sha256:8551cb62a169ec7162ac7be8d4817d561f60e08eaa485234898414bb5a8a0b4c", size = 85196 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/87/f5/72347bc88306acb359581ac4d52f23c0ef445b57157adedb9aee0cd689d2/httpcore-1.0.7-py3-none-any.whl", hash = "sha256:a3fff8f43dc260d5bd363d9f9cf1830fa3a458b332856f34282de498ed420edd", size = 78551 }, +] + +[[package]] +name = "httpx" +version = "0.28.1" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "anyio", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "certifi", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "httpcore", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "idna", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/b1/df/48c586a5fe32a0f01324ee087459e112ebb7224f646c0b5023f5e79e9956/httpx-0.28.1.tar.gz", hash = "sha256:75e98c5f16b0f35b567856f597f06ff2270a374470a5c2392242528e3e3e42fc", size = 141406 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/2a/39/e50c7c3a983047577ee07d2a9e53faf5a69493943ec3f6a384bdc792deb2/httpx-0.28.1-py3-none-any.whl", hash = "sha256:d909fcccc110f8c7faf814ca82a9a4d816bc5a6dbfea25d6591d6985b8ba59ad", size = 73517 }, +] + [[package]] name = "huggingface-hub" version = "0.25.1" @@ -275,6 +721,15 @@ wheels = [ { url = "https://download.pytorch.org/whl/nightly/huggingface_hub-0.25.1-py3-none-any.whl" }, ] +[[package]] +name = "humanfriendly" +version = "10.0" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/cc/3f/2c29224acb2e2df4d2046e4c73ee2662023c58ff5b113c4c1adac0886c43/humanfriendly-10.0.tar.gz", hash = "sha256:6b0b831ce8f15f7300721aa49829fc4e83921a9a301cc7f606be6686a2288ddc", size = 360702 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/f0/0f/310fb31e39e2d734ccaa2c0fb981ee41f7bd5056ce9bc29b2248bd569169/humanfriendly-10.0-py2.py3-none-any.whl", hash = "sha256:1697e1a8a8f550fd43c2865cd84542fc175a61dcb779b6fee18cf6b6ccba1477", size = 86794 }, +] + [[package]] name = "identify" version = "2.6.3" @@ -382,6 +837,55 @@ wheels = [ { url = "https://download.pytorch.org/whl/nightly/jinja2-3.1.4-py3-none-any.whl" }, ] +[[package]] +name = "jiter" +version = "0.8.2" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/f8/70/90bc7bd3932e651486861df5c8ffea4ca7c77d28e8532ddefe2abc561a53/jiter-0.8.2.tar.gz", hash = "sha256:cd73d3e740666d0e639f678adb176fad25c1bcbdae88d8d7b857e1783bb4212d", size = 163007 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/0d/17/57acab00507e60bd954eaec0837d9d7b119b4117ff49b8a62f2b646f32ed/jiter-0.8.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d5c826a221851a8dc028eb6d7d6429ba03184fa3c7e83ae01cd6d3bd1d4bd17d", size = 335465 }, + { url = "https://files.pythonhosted.org/packages/74/b9/1a3ddd2bc95ae17c815b021521020f40c60b32137730126bada962ef32b4/jiter-0.8.2-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:d35c864c2dff13dfd79fb070fc4fc6235d7b9b359efe340e1261deb21b9fcb66", size = 355570 }, + { url = "https://files.pythonhosted.org/packages/78/69/6d29e2296a934199a7d0dde673ecccf98c9c8db44caf0248b3f2b65483cb/jiter-0.8.2-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:f557c55bc2b7676e74d39d19bcb8775ca295c7a028246175d6a8b431e70835e5", size = 381383 }, + { url = "https://files.pythonhosted.org/packages/22/d7/fbc4c3fb1bf65f9be22a32759b539f88e897aeb13fe84ab0266e4423487a/jiter-0.8.2-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:580ccf358539153db147e40751a0b41688a5ceb275e6f3e93d91c9467f42b2e3", size = 390454 }, + { url = "https://files.pythonhosted.org/packages/4d/a0/3993cda2e267fe679b45d0bcc2cef0b4504b0aa810659cdae9737d6bace9/jiter-0.8.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:af102d3372e917cffce49b521e4c32c497515119dc7bd8a75665e90a718bbf08", size = 345039 }, + { url = "https://files.pythonhosted.org/packages/b9/ef/69c18562b4c09ce88fab5df1dcaf643f6b1a8b970b65216e7221169b81c4/jiter-0.8.2-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:cadcc978f82397d515bb2683fc0d50103acff2a180552654bb92d6045dec2c49", size = 376200 }, + { url = "https://files.pythonhosted.org/packages/4d/17/0b5a8de46a6ab4d836f70934036278b49b8530c292b29dde3483326d4555/jiter-0.8.2-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:ba5bdf56969cad2019d4e8ffd3f879b5fdc792624129741d3d83fc832fef8c7d", size = 511158 }, + { url = "https://files.pythonhosted.org/packages/6c/b2/c401a0a2554b36c9e6d6e4876b43790d75139cf3936f0222e675cbc23451/jiter-0.8.2-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:3b94a33a241bee9e34b8481cdcaa3d5c2116f575e0226e421bed3f7a6ea71cff", size = 503956 }, + { url = "https://files.pythonhosted.org/packages/e5/69/64058e18263d9a5f1e10f90c436853616d5f047d997c37c7b2df11b085ec/jiter-0.8.2-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a9584de0cd306072635fe4b89742bf26feae858a0683b399ad0c2509011b9dc0", size = 335506 }, + { url = "https://files.pythonhosted.org/packages/9d/14/b747f9a77b8c0542141d77ca1e2a7523e854754af2c339ac89a8b66527d6/jiter-0.8.2-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:5a90a923338531b7970abb063cfc087eebae6ef8ec8139762007188f6bc69a9f", size = 355849 }, + { url = "https://files.pythonhosted.org/packages/53/e2/98a08161db7cc9d0e39bc385415890928ff09709034982f48eccfca40733/jiter-0.8.2-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d21974d246ed0181558087cd9f76e84e8321091ebfb3a93d4c341479a736f099", size = 381700 }, + { url = "https://files.pythonhosted.org/packages/7a/38/1674672954d35bce3b1c9af99d5849f9256ac8f5b672e020ac7821581206/jiter-0.8.2-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:32475a42b2ea7b344069dc1e81445cfc00b9d0e3ca837f0523072432332e9f74", size = 389710 }, + { url = "https://files.pythonhosted.org/packages/f8/9b/92f9da9a9e107d019bcf883cd9125fa1690079f323f5a9d5c6986eeec3c0/jiter-0.8.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8b9931fd36ee513c26b5bf08c940b0ac875de175341cbdd4fa3be109f0492586", size = 345553 }, + { url = "https://files.pythonhosted.org/packages/44/a6/6d030003394e9659cd0d7136bbeabd82e869849ceccddc34d40abbbbb269/jiter-0.8.2-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:ce0820f4a3a59ddced7fce696d86a096d5cc48d32a4183483a17671a61edfddc", size = 376388 }, + { url = "https://files.pythonhosted.org/packages/ad/8d/87b09e648e4aca5f9af89e3ab3cfb93db2d1e633b2f2931ede8dabd9b19a/jiter-0.8.2-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:8ffc86ae5e3e6a93765d49d1ab47b6075a9c978a2b3b80f0f32628f39caa0c88", size = 511226 }, + { url = "https://files.pythonhosted.org/packages/77/95/8008ebe4cdc82eac1c97864a8042ca7e383ed67e0ec17bfd03797045c727/jiter-0.8.2-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:5127dc1abd809431172bc3fbe8168d6b90556a30bb10acd5ded41c3cfd6f43b6", size = 504134 }, + { url = "https://files.pythonhosted.org/packages/06/99/a2bf660d8ccffee9ad7ed46b4f860d2108a148d0ea36043fd16f4dc37e94/jiter-0.8.2-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:180a8aea058f7535d1c84183c0362c710f4750bef66630c05f40c93c2b152a0f", size = 334242 }, + { url = "https://files.pythonhosted.org/packages/a7/5f/cea1c17864828731f11427b9d1ab7f24764dbd9aaf4648a7f851164d2718/jiter-0.8.2-cp312-cp312-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:025337859077b41548bdcbabe38698bcd93cfe10b06ff66617a48ff92c9aec60", size = 356654 }, + { url = "https://files.pythonhosted.org/packages/e9/13/62774b7e5e7f5d5043efe1d0f94ead66e6d0f894ae010adb56b3f788de71/jiter-0.8.2-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ecff0dc14f409599bbcafa7e470c00b80f17abc14d1405d38ab02e4b42e55b57", size = 379967 }, + { url = "https://files.pythonhosted.org/packages/ec/fb/096b34c553bb0bd3f2289d5013dcad6074948b8d55212aa13a10d44c5326/jiter-0.8.2-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:ffd9fee7d0775ebaba131f7ca2e2d83839a62ad65e8e02fe2bd8fc975cedeb9e", size = 389252 }, + { url = "https://files.pythonhosted.org/packages/17/61/beea645c0bf398ced8b199e377b61eb999d8e46e053bb285c91c3d3eaab0/jiter-0.8.2-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:14601dcac4889e0a1c75ccf6a0e4baf70dbc75041e51bcf8d0e9274519df6887", size = 345490 }, + { url = "https://files.pythonhosted.org/packages/d5/df/834aa17ad5dcc3cf0118821da0a0cf1589ea7db9832589278553640366bc/jiter-0.8.2-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:92249669925bc1c54fcd2ec73f70f2c1d6a817928480ee1c65af5f6b81cdf12d", size = 376991 }, + { url = "https://files.pythonhosted.org/packages/67/80/87d140399d382fb4ea5b3d56e7ecaa4efdca17cd7411ff904c1517855314/jiter-0.8.2-cp312-cp312-musllinux_1_1_aarch64.whl", hash = "sha256:e725edd0929fa79f8349ab4ec7f81c714df51dc4e991539a578e5018fa4a7152", size = 510822 }, + { url = "https://files.pythonhosted.org/packages/5c/37/3394bb47bac1ad2cb0465601f86828a0518d07828a650722e55268cdb7e6/jiter-0.8.2-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:bf55846c7b7a680eebaf9c3c48d630e1bf51bdf76c68a5f654b8524335b0ad29", size = 503730 }, + { url = "https://files.pythonhosted.org/packages/7f/68/805978f2f446fa6362ba0cc2e4489b945695940656edd844e110a61c98f8/jiter-0.8.2-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:58dc9bc9767a1101f4e5e22db1b652161a225874d66f0e5cb8e2c7d1c438b587", size = 333918 }, + { url = "https://files.pythonhosted.org/packages/b3/99/0f71f7be667c33403fa9706e5b50583ae5106d96fab997fa7e2f38ee8347/jiter-0.8.2-cp313-cp313-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:37b2998606d6dadbb5ccda959a33d6a5e853252d921fec1792fc902351bb4e2c", size = 356057 }, + { url = "https://files.pythonhosted.org/packages/8d/50/a82796e421a22b699ee4d2ce527e5bcb29471a2351cbdc931819d941a167/jiter-0.8.2-cp313-cp313-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4ab9a87f3784eb0e098f84a32670cfe4a79cb6512fd8f42ae3d0709f06405d18", size = 379790 }, + { url = "https://files.pythonhosted.org/packages/3c/31/10fb012b00f6d83342ca9e2c9618869ab449f1aa78c8f1b2193a6b49647c/jiter-0.8.2-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:79aec8172b9e3c6d05fd4b219d5de1ac616bd8da934107325a6c0d0e866a21b6", size = 388285 }, + { url = "https://files.pythonhosted.org/packages/c8/81/f15ebf7de57be488aa22944bf4274962aca8092e4f7817f92ffa50d3ee46/jiter-0.8.2-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:711e408732d4e9a0208008e5892c2966b485c783cd2d9a681f3eb147cf36c7ef", size = 344764 }, + { url = "https://files.pythonhosted.org/packages/b3/e8/0cae550d72b48829ba653eb348cdc25f3f06f8a62363723702ec18e7be9c/jiter-0.8.2-cp313-cp313-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:653cf462db4e8c41995e33d865965e79641ef45369d8a11f54cd30888b7e6ff1", size = 376620 }, + { url = "https://files.pythonhosted.org/packages/b8/50/e5478ff9d82534a944c03b63bc217c5f37019d4a34d288db0f079b13c10b/jiter-0.8.2-cp313-cp313-musllinux_1_1_aarch64.whl", hash = "sha256:9c63eaef32b7bebac8ebebf4dabebdbc6769a09c127294db6babee38e9f405b9", size = 510402 }, + { url = "https://files.pythonhosted.org/packages/8e/1e/3de48bbebbc8f7025bd454cedc8c62378c0e32dd483dece5f4a814a5cb55/jiter-0.8.2-cp313-cp313-musllinux_1_1_x86_64.whl", hash = "sha256:eb21aaa9a200d0a80dacc7a81038d2e476ffe473ffdd9c91eb745d623561de05", size = 503018 }, + { url = "https://files.pythonhosted.org/packages/a0/4c/c02408042e6a7605ec063daed138e07b982fdb98467deaaf1c90950cf2c6/jiter-0.8.2-cp313-cp313t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b2dd880785088ff2ad21ffee205e58a8c1ddabc63612444ae41e5e4b321b39c0", size = 342875 }, + { url = "https://files.pythonhosted.org/packages/56/4c/b413977c20bbb359b4d6c91d04f7f36fc525af0b7778119815477fc97242/jiter-0.8.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f7200b8f7619d36aa51c803fd52020a2dfbea36ffec1b5e22cab11fd34d95a6d", size = 335344 }, + { url = "https://files.pythonhosted.org/packages/b0/59/51b080519938192edd33b4e8d48adb7e9bf9e0d699ec8b91119b9269fc75/jiter-0.8.2-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:70bf4c43652cc294040dbb62256c83c8718370c8b93dd93d934b9a7bf6c4f53c", size = 356298 }, + { url = "https://files.pythonhosted.org/packages/72/bb/828db5ea406916d7b2232be31393f782b0f71bcb0b128750c4a028157565/jiter-0.8.2-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:f9d471356dc16f84ed48768b8ee79f29514295c7295cb41e1133ec0b2b8d637d", size = 381703 }, + { url = "https://files.pythonhosted.org/packages/c0/88/45d33a8728733e161e9783c54d8ecca0fc4c1aa74b1cebea1d97917eddc3/jiter-0.8.2-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:859e8eb3507894093d01929e12e267f83b1d5f6221099d3ec976f0c995cb6bd9", size = 391281 }, + { url = "https://files.pythonhosted.org/packages/45/3e/142712e0f45c28ad8a678dc8732a78294ce5a36fc694141f772bb827a8f2/jiter-0.8.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:eaa58399c01db555346647a907b4ef6d4f584b123943be6ed5588c3f2359c9f4", size = 345553 }, + { url = "https://files.pythonhosted.org/packages/36/42/9b463b59fd22687b6da1afcad6c9adc870464a808208651de73f1dbeda09/jiter-0.8.2-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:8f2d5ed877f089862f4c7aacf3a542627c1496f972a34d0474ce85ee7d939c27", size = 377063 }, + { url = "https://files.pythonhosted.org/packages/83/b3/44b1f5cd2e4eb15757eec341b25399da4c90515bb881ef6636b50a8c08a5/jiter-0.8.2-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:03c9df035d4f8d647f8c210ddc2ae0728387275340668fb30d2421e17d9a0841", size = 512543 }, + { url = "https://files.pythonhosted.org/packages/46/4e/c695c803aa2b668c057b2dea1cdd7a884d1a819ce610cec0be9666210bfd/jiter-0.8.2-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:8bd2a824d08d8977bb2794ea2682f898ad3d8837932e3a74937e93d62ecbb637", size = 505141 }, +] + [[package]] name = "jupyterlab-widgets" version = "3.0.13" @@ -391,6 +895,15 @@ wheels = [ { url = "https://files.pythonhosted.org/packages/a9/93/858e87edc634d628e5d752ba944c2833133a28fa87bb093e6832ced36a3e/jupyterlab_widgets-3.0.13-py3-none-any.whl", hash = "sha256:e3cda2c233ce144192f1e29914ad522b2f4c40e77214b0cc97377ca3d323db54", size = 214392 }, ] +[[package]] +name = "lark" +version = "1.2.2" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/af/60/bc7622aefb2aee1c0b4ba23c1446d3e30225c8770b38d7aedbfb65ca9d5a/lark-1.2.2.tar.gz", hash = "sha256:ca807d0162cd16cef15a8feecb862d7319e7a09bdb13aef927968e45040fed80", size = 252132 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/2d/00/d90b10b962b4277f5e64a78b6609968859ff86889f5b898c1a778c06ec00/lark-1.2.2-py3-none-any.whl", hash = "sha256:c2276486b02f0f1b90be155f2c8ba4a8e194d42775786db622faccd652d8e80c", size = 111036 }, +] + [[package]] name = "markdown-it-py" version = "3.0.0" @@ -444,6 +957,12 @@ wheels = [ { url = "https://files.pythonhosted.org/packages/b3/38/89ba8ad64ae25be8de66a6d463314cf1eb366222074cfda9ee839c56a4b4/mdurl-0.1.2-py3-none-any.whl", hash = "sha256:84008a41e51615a49fc9966191ff91509e3c40b939176e643fd50a5c2196b8f8", size = 9979 }, ] +[[package]] +name = "mpi4py" +version = "4.0.1" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/08/34/8499a92a387d24d0092c38089f8195f13c5c76f0f814126af3fe363e5636/mpi4py-4.0.1.tar.gz", hash = "sha256:f3174b245775d556f4fddb32519a2066ef0592edc810c5b5a59238f9a0a40c89", size = 466179 } + [[package]] name = "mpmath" version = "1.3.0" @@ -452,28 +971,82 @@ wheels = [ { url = "https://download.pytorch.org/whl/nightly/mpmath-1.3.0-py3-none-any.whl" }, ] +[[package]] +name = "multidict" +version = "6.1.0" +source = { registry = "https://download.pytorch.org/whl/nightly/cu126" } +dependencies = [ + { name = "typing-extensions", marker = "(python_full_version < '3.11' and sys_platform == 'linux') or (python_full_version < '3.11' and sys_platform == 'windows')" }, +] +wheels = [ + { url = "https://download.pytorch.org/whl/nightly/multidict-6.1.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/multidict-6.1.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl" }, + { url = "https://download.pytorch.org/whl/nightly/multidict-6.1.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl" }, + { url = "https://download.pytorch.org/whl/nightly/multidict-6.1.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/multidict-6.1.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl" }, + { url = "https://download.pytorch.org/whl/nightly/multidict-6.1.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/multidict-6.1.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl" }, + { url = "https://download.pytorch.org/whl/nightly/multidict-6.1.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl" }, + { url = "https://download.pytorch.org/whl/nightly/multidict-6.1.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/multidict-6.1.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl" }, + { url = "https://download.pytorch.org/whl/nightly/multidict-6.1.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/multidict-6.1.0-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl" }, + { url = "https://download.pytorch.org/whl/nightly/multidict-6.1.0-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl" }, + { url = "https://download.pytorch.org/whl/nightly/multidict-6.1.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/multidict-6.1.0-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl" }, + { url = "https://download.pytorch.org/whl/nightly/multidict-6.1.0-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/multidict-6.1.0-cp313-cp313-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl" }, + { url = "https://download.pytorch.org/whl/nightly/multidict-6.1.0-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl" }, + { url = "https://download.pytorch.org/whl/nightly/multidict-6.1.0-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/multidict-6.1.0-cp313-cp313-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl" }, + { url = "https://download.pytorch.org/whl/nightly/multidict-6.1.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/multidict-6.1.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl" }, + { url = "https://download.pytorch.org/whl/nightly/multidict-6.1.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl" }, + { url = "https://download.pytorch.org/whl/nightly/multidict-6.1.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/multidict-6.1.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl" }, + { url = "https://download.pytorch.org/whl/nightly/multidict-6.1.0-py3-none-any.whl" }, +] + +[[package]] +name = "multiprocess" +version = "0.70.15" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "dill", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/68/e0/a77ca96e772e13c828fa52f3ad370d413bef194aeaf78b7c6611870ad815/multiprocess-0.70.15.tar.gz", hash = "sha256:f20eed3036c0ef477b07a4177cf7c1ba520d9a2677870a4f47fe026f0cd6787e", size = 1894495 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/83/97/08402e5ec72c1cc461df1f2e654c1e55527bb05c22120dcdf59745189df6/multiprocess-0.70.15-pp39-pypy39_pp73-manylinux_2_24_i686.whl", hash = "sha256:0eac53214d664c49a34695e5824872db4006b1a465edd7459a251809c3773370", size = 133501 }, + { url = "https://files.pythonhosted.org/packages/71/47/5d12db2427465486c0b336cc67753e8826b756a2e4d4ef6385f4f01b355b/multiprocess-0.70.15-pp39-pypy39_pp73-manylinux_2_24_x86_64.whl", hash = "sha256:1a51dd34096db47fb21fa2b839e615b051d51b97af9a67afbcdaa67186b44883", size = 133501 }, + { url = "https://files.pythonhosted.org/packages/35/a8/36d8d7b3e46b377800d8dec47891cdf05842d1a2366909ae4a0c89fbc5e6/multiprocess-0.70.15-py310-none-any.whl", hash = "sha256:7dd58e33235e83cf09d625e55cffd7b0f0eede7ee9223cdd666a87624f60c21a", size = 134824 }, + { url = "https://files.pythonhosted.org/packages/e7/41/96ac938770ba6e7d5ae1d8c9cafebac54b413549042c6260f0d0a6ec6622/multiprocess-0.70.15-py311-none-any.whl", hash = "sha256:134f89053d82c9ed3b73edd3a2531eb791e602d4f4156fc92a79259590bd9670", size = 135392 }, + { url = "https://files.pythonhosted.org/packages/ca/3f/8354ce12fd13bd5c5bb4722261a10ca1d6e2eb7c1c08fa3d8a4e9dc98f44/multiprocess-0.70.15-py37-none-any.whl", hash = "sha256:f7d4a1629bccb433114c3b4885f69eccc200994323c80f6feee73b0edc9199c5", size = 116276 }, + { url = "https://files.pythonhosted.org/packages/c2/a6/c5cb599d917904878f220a4dbdfdcc4ef291dd3956c35b3b0dc6fc42fb6d/multiprocess-0.70.15-py38-none-any.whl", hash = "sha256:bee9afba476c91f9ebee7beeee0601face9eff67d822e893f9a893725fbd6316", size = 132626 }, + { url = "https://files.pythonhosted.org/packages/c6/c9/820b5ab056f4ada76fbe05bd481a948f287957d6cbfd59e2dd2618b408c1/multiprocess-0.70.15-py39-none-any.whl", hash = "sha256:3e0953f5d52b4c76f1c973eaf8214554d146f2be5decb48e928e55c7a2d19338", size = 133349 }, +] + [[package]] name = "mypy" -version = "1.13.0" +version = "1.14.0" source = { registry = "https://pypi.org/simple" } dependencies = [ { name = "mypy-extensions", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, { name = "tomli", marker = "(python_full_version < '3.11' and sys_platform == 'linux') or (python_full_version < '3.11' and sys_platform == 'windows')" }, { name = "typing-extensions", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, ] -sdist = { url = "https://files.pythonhosted.org/packages/e8/21/7e9e523537991d145ab8a0a2fd98548d67646dc2aaaf6091c31ad883e7c1/mypy-1.13.0.tar.gz", hash = "sha256:0291a61b6fbf3e6673e3405cfcc0e7650bebc7939659fdca2702958038bd835e", size = 3152532 } +sdist = { url = "https://files.pythonhosted.org/packages/8c/7b/08046ef9330735f536a09a2e31b00f42bccdb2795dcd979636ba43bb2d63/mypy-1.14.0.tar.gz", hash = "sha256:822dbd184d4a9804df5a7d5335a68cf7662930e70b8c1bc976645d1509f9a9d6", size = 3215684 } wheels = [ - { url = "https://files.pythonhosted.org/packages/a5/2d/4a23849729bb27934a0e079c9c1aad912167d875c7b070382a408d459651/mypy-1.13.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:7b2353a44d2179846a096e25691d54d59904559f4232519d420d64da6828a3a7", size = 12587706 }, - { url = "https://files.pythonhosted.org/packages/5c/c3/d318e38ada50255e22e23353a469c791379825240e71b0ad03e76ca07ae6/mypy-1.13.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:0730d1c6a2739d4511dc4253f8274cdd140c55c32dfb0a4cf8b7a43f40abfa6f", size = 13105586 }, - { url = "https://files.pythonhosted.org/packages/26/50/29d3e7dd166e74dc13d46050b23f7d6d7533acf48f5217663a3719db024e/mypy-1.13.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:20c7ee0bc0d5a9595c46f38beb04201f2620065a93755704e141fcac9f59db2b", size = 12506263 }, - { url = "https://files.pythonhosted.org/packages/3f/1d/676e76f07f7d5ddcd4227af3938a9c9640f293b7d8a44dd4ff41d4db25c1/mypy-1.13.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:3790ded76f0b34bc9c8ba4def8f919dd6a46db0f5a6610fb994fe8efdd447f73", size = 12984688 }, - { url = "https://files.pythonhosted.org/packages/ba/07/37d67048786ae84e6612575e173d713c9a05d0ae495dde1e68d972207d98/mypy-1.13.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:164f28cb9d6367439031f4c81e84d3ccaa1e19232d9d05d37cb0bd880d3f93c2", size = 12589275 }, - { url = "https://files.pythonhosted.org/packages/1f/17/b1018c6bb3e9f1ce3956722b3bf91bff86c1cefccca71cec05eae49d6d41/mypy-1.13.0-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:a4c1bfcdbce96ff5d96fc9b08e3831acb30dc44ab02671eca5953eadad07d6d0", size = 13037783 }, - { url = "https://files.pythonhosted.org/packages/d1/1f/6b76be289a5a521bb1caedc1f08e76ff17ab59061007f201a8a18cc514d1/mypy-1.13.0-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:de2904956dac40ced10931ac967ae63c5089bd498542194b436eb097a9f77bc8", size = 12584043 }, - { url = "https://files.pythonhosted.org/packages/a6/83/5a85c9a5976c6f96e3a5a7591aa28b4a6ca3a07e9e5ba0cec090c8b596d6/mypy-1.13.0-cp313-cp313-musllinux_1_1_x86_64.whl", hash = "sha256:7bfd8836970d33c2105562650656b6846149374dc8ed77d98424b40b09340ba7", size = 13036996 }, - { url = "https://files.pythonhosted.org/packages/38/e9/fc3865e417722f98d58409770be01afb961e2c1f99930659ff4ae7ca8b7e/mypy-1.13.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:7029881ec6ffb8bc233a4fa364736789582c738217b133f1b55967115288a2bc", size = 12586394 }, - { url = "https://files.pythonhosted.org/packages/2e/35/f4d8b6d2cb0b3dad63e96caf159419dda023f45a358c6c9ac582ccaee354/mypy-1.13.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:3e38b980e5681f28f033f3be86b099a247b13c491f14bb8b1e1e134d23bb599d", size = 13103591 }, - { url = "https://files.pythonhosted.org/packages/3b/86/72ce7f57431d87a7ff17d442f521146a6585019eb8f4f31b7c02801f78ad/mypy-1.13.0-py3-none-any.whl", hash = "sha256:9c250883f9fd81d212e0952c92dbfcc96fc237f4b7c92f56ac81fd48460b3e5a", size = 2647043 }, + { url = "https://files.pythonhosted.org/packages/f5/00/56b1619ff1f3fcad2d411eccda60d74d20e73bda39c218d5ad2769980682/mypy-1.14.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:1628c5c3ce823d296e41e2984ff88c5861499041cb416a8809615d0c1f41740e", size = 12832976 }, + { url = "https://files.pythonhosted.org/packages/e7/8b/9247838774b0bd865f190cc221822212091317f16310305ef924d9772532/mypy-1.14.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:7fadb29b77fc14a0dd81304ed73c828c3e5cde0016c7e668a86a3e0dfc9f3af3", size = 13013704 }, + { url = "https://files.pythonhosted.org/packages/69/2c/3dbe51877a24daa467f8d8631f9ffd1aabbf0f6d9367a01c44a59df81fe0/mypy-1.14.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:1daca283d732943731a6a9f20fdbcaa927f160bc51602b1d4ef880a6fb252015", size = 12746528 }, + { url = "https://files.pythonhosted.org/packages/a1/a8/eb20cde4ba9c4c3e20d958918a7c5d92210f4d1a0200c27de9a641f70996/mypy-1.14.0-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:7e68047bedb04c1c25bba9901ea46ff60d5eaac2d71b1f2161f33107e2b368eb", size = 12883489 }, + { url = "https://files.pythonhosted.org/packages/3e/38/7db2c5d0f4d290e998f7a52b2e2616c7bbad96b8e04278ab09d11978a29e/mypy-1.14.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:390dfb898239c25289495500f12fa73aa7f24a4c6d90ccdc165762462b998d63", size = 12862786 }, + { url = "https://files.pythonhosted.org/packages/bf/4b/62d59c801b34141040989949c2b5c157d0408b45357335d3ec5b2845b0f6/mypy-1.14.0-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:7e026d55ddcd76e29e87865c08cbe2d0104e2b3153a523c529de584759379d3d", size = 12971568 }, + { url = "https://files.pythonhosted.org/packages/8b/cf/7a8ae5c0161edae15d25c2c67c68ce8b150cbdc45aefc13a8be271ee80b2/mypy-1.14.0-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:342de51c48bab326bfc77ce056ba08c076d82ce4f5a86621f972ed39970f94d8", size = 12867676 }, + { url = "https://files.pythonhosted.org/packages/9c/d0/71f7bbdcc7cfd0f2892db5b13b1e8857673f2cc9e0c30e3e4340523dc186/mypy-1.14.0-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:00df23b42e533e02a6f0055e54de9a6ed491cd8b7ea738647364fd3a39ea7efc", size = 12964189 }, + { url = "https://files.pythonhosted.org/packages/bb/bd/a0eb1789dfeaab0ca93d00f373c002cac4734e8f902de4e7eceb9245a116/mypy-1.14.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:f0b343a1d3989547024377c2ba0dca9c74a2428ad6ed24283c213af8dbb0710b", size = 12832863 }, + { url = "https://files.pythonhosted.org/packages/ef/0c/d404be19b1145f9371c4d4fdfc166337a2810c2be0f19dec5965186e8fab/mypy-1.14.0-cp39-cp39-musllinux_1_2_x86_64.whl", hash = "sha256:cdb5563c1726c85fb201be383168f8c866032db95e1095600806625b3a648cb7", size = 13008982 }, + { url = "https://files.pythonhosted.org/packages/39/32/0214608af400cdf8f5102144bb8af10d880675c65ed0b58f7e0e77175d50/mypy-1.14.0-py3-none-any.whl", hash = "sha256:2238d7f93fc4027ed1efc944507683df3ba406445a2b6c96e79666a045aadfab", size = 2752803 }, ] [[package]] @@ -494,22 +1067,22 @@ wheels = [ [[package]] name = "ninja" -version = "1.11.1.2" +version = "1.11.1.3" source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/b4/49/4f1a79f99f4c3eb5d22f943bba14832923bb44423254d5089d38a9f6da63/ninja-1.11.1.2.tar.gz", hash = "sha256:4fbd07b2b4232543726abafdd350453a2fabef4527664ca0e491c578aee5f857", size = 129009 } +sdist = { url = "https://files.pythonhosted.org/packages/bd/8f/21a2701f95b7d0d5137736561b3427ece0c4a1e085d4a223b92d16ab7d8b/ninja-1.11.1.3.tar.gz", hash = "sha256:edfa0d2e9d7ead1635b03e40a32ad56cc8f56798b6e2e9848d8300b174897076", size = 129532 } wheels = [ - { url = "https://files.pythonhosted.org/packages/7b/87/d33b00c6168915b343fde8877a6852692ba6f7d3ebee07f251a2dc338563/ninja-1.11.1.2-py3-none-manylinux_2_12_i686.manylinux2010_i686.whl", hash = "sha256:ab4068ff7ff1f895485ad604116165b05d6810c802170a7f22c09dd678d5587d", size = 472101 }, - { url = "https://files.pythonhosted.org/packages/62/54/787bb70e6af2f1b1853af9bab62a5e7cb35b957d72daf253b7f3c653c005/ninja-1.11.1.2-py3-none-manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:33d258809c8eda81f9d80e18a081a6eef3215e5fd1ba8902400d786641994e89", size = 422889 }, - { url = "https://files.pythonhosted.org/packages/27/9f/1a021b766134f4ea91346fbbf7653e17a483242929c9c579b769830bdcd6/ninja-1.11.1.2-py3-none-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ed25892c16e49e66383a8db6a67a9f33b41230fc485426094d7da51e2255ec2b", size = 157046 }, - { url = "https://files.pythonhosted.org/packages/a8/e3/e05286d374e69272bd0a00517f76effe026207cb07a9d269cc3abdfe4bdd/ninja-1.11.1.2-py3-none-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:232767144401847db62e8392047866698bb3678158a1ae4400a97111110e90f2", size = 180014 }, - { url = "https://files.pythonhosted.org/packages/22/b4/0fb29155c05685a8a4d20489b90c340dd781db5c14b5586075fcbdf748e4/ninja-1.11.1.2-py3-none-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9200247cf4c1643a67d079836b8dd31a362e34e618b50b5e3a5c0d0171efc442", size = 157099 }, - { url = "https://files.pythonhosted.org/packages/e7/85/d67805c3d47c902f7b1a1a5b75317f4d45af7bb7132c342adf47eafc66b8/ninja-1.11.1.2-py3-none-manylinux_2_28_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:0c9c36f6e6f8946c7271b0ed14d98fc3ea467a0c0954fb73f5f656c42667d943", size = 130093 }, - { url = "https://files.pythonhosted.org/packages/eb/40/9a7fc0e417b1aab20f91be957418d2e5952db9f9b72f4396a8a097310964/ninja-1.11.1.2-py3-none-musllinux_1_1_aarch64.whl", hash = "sha256:3e815e4147832b17ec38417efcb31df51671ae273f083409304c7cc32a14dd1a", size = 372508 }, - { url = "https://files.pythonhosted.org/packages/0e/db/8c4843e8454e1ec2e6651b5caef31762e46fbaf3a71e6817e7d9cd28b5cb/ninja-1.11.1.2-py3-none-musllinux_1_1_i686.whl", hash = "sha256:ecf3df324b56fdfb0872990a71e706efdae286e010310816c72b6bf24431711b", size = 419368 }, - { url = "https://files.pythonhosted.org/packages/c3/e0/17ccb830c1638966d75a19a59e0ce55aadb4cf5c2cae5bcf97f74511c33e/ninja-1.11.1.2-py3-none-musllinux_1_1_ppc64le.whl", hash = "sha256:cb6b476eb4e84c0efcfd3ab04f660dedce8adb854b56b043639312f3af176df6", size = 420305 }, - { url = "https://files.pythonhosted.org/packages/30/e4/7d02c7a633c36a9aa7433fb742931a62f0a3aa72b484ed23d73cc6415286/ninja-1.11.1.2-py3-none-musllinux_1_1_s390x.whl", hash = "sha256:508fb93395a5c82a4d99d30fce0cbaf5cb2bd33e5c1dc9faaa080e199802dbc9", size = 416060 }, - { url = "https://files.pythonhosted.org/packages/0d/11/4dc053f20c64f5a340d72f948bbad22818d242afd54e826e0c95ca3779fe/ninja-1.11.1.2-py3-none-musllinux_1_1_x86_64.whl", hash = "sha256:52af7f45750c5c288d566fd0c927ed9bb0d8f2e50803709f582a42bcc4ec167b", size = 379729 }, - { url = "https://files.pythonhosted.org/packages/ab/57/adaa8052ae4854c5f8e228baa1a77aad68093bc1aedf32597fa5e7714118/ninja-1.11.1.2-py3-none-musllinux_1_2_armv7l.whl", hash = "sha256:99fc4b87299242e10d7edd1c7737fdfb1269019e32f9f4267630887f6183a49e", size = 434886 }, + { url = "https://files.pythonhosted.org/packages/72/6b/3805be87df8417a0c7b21078c8045f2a1e59b34f371bfe4cb4fb0d6df7f2/ninja-1.11.1.3-py3-none-manylinux_2_12_i686.manylinux2010_i686.whl", hash = "sha256:bc3ebc8b2e47716149f3541742b5cd8e0b08f51013b825c05baca3e34854370d", size = 472101 }, + { url = "https://files.pythonhosted.org/packages/6b/35/a8e38d54768e67324e365e2a41162be298f51ec93e6bd4b18d237d7250d8/ninja-1.11.1.3-py3-none-manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:a27e78ca71316c8654965ee94b286a98c83877bfebe2607db96897bbfe458af0", size = 422884 }, + { url = "https://files.pythonhosted.org/packages/2f/99/7996457319e139c02697fb2aa28e42fe32bb0752cef492edc69d56a3552e/ninja-1.11.1.3-py3-none-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:2883ea46b3c5079074f56820f9989c6261fcc6fd873d914ee49010ecf283c3b2", size = 157046 }, + { url = "https://files.pythonhosted.org/packages/6d/8b/93f38e5cddf76ccfdab70946515b554f25d2b4c95ef9b2f9cfbc43fa7cc1/ninja-1.11.1.3-py3-none-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:8c4bdb9fd2d0c06501ae15abfd23407660e95659e384acd36e013b6dd7d8a8e4", size = 180014 }, + { url = "https://files.pythonhosted.org/packages/7d/1d/713884d0fa3c972164f69d552e0701d30e2bf25eba9ef160bfb3dc69926a/ninja-1.11.1.3-py3-none-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:114ed5c61c8474df6a69ab89097a20749b769e2c219a452cb2fadc49b0d581b0", size = 157098 }, + { url = "https://files.pythonhosted.org/packages/c7/22/ecb0f70e77c9e22ee250aa717a608a142756833a34d43943d7d658ee0e56/ninja-1.11.1.3-py3-none-manylinux_2_28_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:7fa2247fce98f683bc712562d82b22b8a0a5c000738a13147ca2d1b68c122298", size = 130089 }, + { url = "https://files.pythonhosted.org/packages/ec/a6/3ee846c20ab6ad95b90c5c8703c76cb1f39cc8ce2d1ae468956e3b1b2581/ninja-1.11.1.3-py3-none-musllinux_1_1_aarch64.whl", hash = "sha256:a38c6c6c8032bed68b70c3b065d944c35e9f903342875d3a3218c1607987077c", size = 372508 }, + { url = "https://files.pythonhosted.org/packages/95/0d/aa44abe4141f29148ce671ac8c92045878906b18691c6f87a29711c2ff1c/ninja-1.11.1.3-py3-none-musllinux_1_1_i686.whl", hash = "sha256:56ada5d33b8741d298836644042faddebc83ee669782d661e21563034beb5aba", size = 419369 }, + { url = "https://files.pythonhosted.org/packages/f7/ec/48bf5105568ac9bd2016b701777bdd5000cc09a14ac837fef9f15e8d634e/ninja-1.11.1.3-py3-none-musllinux_1_1_ppc64le.whl", hash = "sha256:53409151da081f3c198bb0bfc220a7f4e821e022c5b7d29719adda892ddb31bb", size = 420304 }, + { url = "https://files.pythonhosted.org/packages/18/e5/69df63976cf971a03379899f8520a036c9dbab26330b37197512aed5b3df/ninja-1.11.1.3-py3-none-musllinux_1_1_s390x.whl", hash = "sha256:1ad2112c2b0159ed7c4ae3731595191b1546ba62316fc40808edecd0306fefa3", size = 416056 }, + { url = "https://files.pythonhosted.org/packages/6f/4f/bdb401af7ed0e24a3fef058e13a149f2de1ce4b176699076993615d55610/ninja-1.11.1.3-py3-none-musllinux_1_1_x86_64.whl", hash = "sha256:28aea3c1c280cba95b8608d50797169f3a34280e3e9a6379b6e340f0c9eaeeb0", size = 379725 }, + { url = "https://files.pythonhosted.org/packages/bd/68/05e7863bf13128c61652eeb3ec7096c3d3a602f32f31752dbfb034e3fa07/ninja-1.11.1.3-py3-none-musllinux_1_2_armv7l.whl", hash = "sha256:b6966f83064a88a51693073eea3decd47e08c3965241e09578ef7aa3a7738329", size = 434881 }, ] [[package]] @@ -633,9 +1206,18 @@ wheels = [ { url = "https://download.pytorch.org/whl/nightly/cu126/nvidia_cusparselt_cu12-0.6.3-py3-none-manylinux2014_x86_64.whl" }, ] +[[package]] +name = "nvidia-ml-py" +version = "12.560.30" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/53/10/5f34de4a71db8b2b7ec4269f4a33287f24c23e2857ea3187c977b7bc3604/nvidia-ml-py-12.560.30.tar.gz", hash = "sha256:f0254dc7400647680a072ee02509bfd46102b60bdfeca321576d4d4817e7fe97", size = 39194 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/b7/f3/a69ce0b1a1e12fbf6b2ad9f4c14c9999fdbdf15f2478d210f0fd501ddc98/nvidia_ml_py-12.560.30-py3-none-any.whl", hash = "sha256:fea371c94d63e38a611c17bbb85fe400e9c8ddb9e8684a9cd0e47786a4bc3c73", size = 40526 }, +] + [[package]] name = "nvidia-modelopt" -version = "0.17.0" +version = "0.21.1" source = { registry = "https://pypi.org/simple" } dependencies = [ { name = "cloudpickle", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, @@ -649,14 +1231,14 @@ dependencies = [ { name = "tqdm", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, ] wheels = [ - { url = "https://files.pythonhosted.org/packages/32/b2/2a688cc56d875a08e3e732af642d0ae0f4a7253dc1a00fd271e2fe1a79e9/nvidia_modelopt-0.17.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f646da43a46ddf10eb2c2ddebe49cb1a58e631348808c5640afe217f8ab223ae", size = 4699528 }, - { url = "https://files.pythonhosted.org/packages/b2/b3/98bea42c27fb9ca9f6c502ec3624f2000732f9cc642652b75378f6fed1db/nvidia_modelopt-0.17.0-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:67cd2faad4c0084864330533112cb231fdf404526086f0803e593f65b7868f47", size = 4502680 }, - { url = "https://files.pythonhosted.org/packages/26/7e/beb7461b6bedf7fff043ded616a520468c992df9d005d3bfd87eab1dab71/nvidia_modelopt-0.17.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ee7480fdb7e1e6d22e17092bfbc6abf6140f070853c21662e4b6162aec228feb", size = 5108601 }, - { url = "https://files.pythonhosted.org/packages/43/79/45572053e3e928f32d62fcae8704a2667a646356ac752a15490829398987/nvidia_modelopt-0.17.0-cp311-cp311-manylinux_2_28_aarch64.whl", hash = "sha256:16e6a8df04e6551a9f2059cb5c68608d18b38cb14dec609ac920035ceeccbf98", size = 4966637 }, - { url = "https://files.pythonhosted.org/packages/7e/a9/53a82e52fd0d5c44a6a09f48db30b7e133e393ad1b7cef280ac41a4a8bca/nvidia_modelopt-0.17.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:557b92dbeeb4d9dedb58d9a6251901870a7458f6e773838efa9155693c5d21e1", size = 5145613 }, - { url = "https://files.pythonhosted.org/packages/4e/59/d3d2c7d59c5b56c86c0c1c55896d6800539f5b1aae7c3ea43331c117fe95/nvidia_modelopt-0.17.0-cp312-cp312-manylinux_2_28_aarch64.whl", hash = "sha256:3a07969bb8c684b3e9f9eecc018bc465d0f683d9e9a4792b5eb435b9ad85f4a5", size = 4949031 }, - { url = "https://files.pythonhosted.org/packages/61/00/39c8ad3969003f7ba87a54197626aa822047bd98a0e09428045e9dd00335/nvidia_modelopt-0.17.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0da51c4464af883eb722b20679633aabc31fadff6b4bee3c33eb4eda3d920484", size = 4694964 }, - { url = "https://files.pythonhosted.org/packages/41/6b/7eb1d60a4706ee1346990bb7b45bf4244be60e24c52894fd817f5ffb649a/nvidia_modelopt-0.17.0-cp39-cp39-manylinux_2_28_aarch64.whl", hash = "sha256:fdc86a37698cad0d780e91ac07166ef4e3b073e87200116351f0b2c7dcada2bf", size = 4502582 }, + { url = "https://files.pythonhosted.org/packages/ef/69/a49a427ae0839a3cdc5eb1868caf563d74d014d9ec4aa8475e3cf18cb7c6/nvidia_modelopt-0.21.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8fef6a2ea3bf51a62c31226825d7d35f883e8d260f4d57a7fcb3b6209431351e", size = 2502694 }, + { url = "https://files.pythonhosted.org/packages/cd/7d/1cd0b8001924e48d54f26c3a92d4afc9340af44e4282090a3f166ebf4943/nvidia_modelopt-0.21.1-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:0d26c3193694f90e8a940133d559d3ce44045a4544b131dd7f29f718110413ca", size = 2590928 }, + { url = "https://files.pythonhosted.org/packages/64/85/2531c5ad2881625fda1f5c4e7def6f53b58b7266f2137e94157cab5f2ea5/nvidia_modelopt-0.21.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3b4410bab11e7c12ac5fdbb1a11a8d485b65e85867496ad91e849e449e27a726", size = 2710695 }, + { url = "https://files.pythonhosted.org/packages/a0/95/774371981a2f564dc4855c18d21f861084ceb686bac751ef6a0d1b03993c/nvidia_modelopt-0.21.1-cp311-cp311-manylinux_2_28_aarch64.whl", hash = "sha256:1bab36c96598615f81efef4be148a23460c4a321289b5b789e33c6f179fdd932", size = 2816194 }, + { url = "https://files.pythonhosted.org/packages/fc/df/55326825e6c870a7d9af5417f734d0d0c1bb1277bb6b352889a4514cff5a/nvidia_modelopt-0.21.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f29fb9d39e6ba51e0f546a93ad9b275d888bd22f74c37547783a4708b9309821", size = 2715861 }, + { url = "https://files.pythonhosted.org/packages/3b/4a/955bc90607e944a8d08a5f77fe026a1102412195ae5e921bdfb4dda14dd7/nvidia_modelopt-0.21.1-cp312-cp312-manylinux_2_28_aarch64.whl", hash = "sha256:d2b8656057ed0d7505d4895542ecdaa6df7acbb6cebb6ff06182450f5ac5d570", size = 2813404 }, + { url = "https://files.pythonhosted.org/packages/39/ed/b5855c053c96990f7d38113c9694df4c42130710abb0eba7d0a3bef793c9/nvidia_modelopt-0.21.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9dfcd53a4b46b4e6ad2744aa389b1dd68560d748f1645b4facf563e3b5b79a05", size = 2498183 }, + { url = "https://files.pythonhosted.org/packages/f5/fd/2b99bbb2d3bb714ac94d35fa32795f2c70814e3cc3aed1612cea41ca10de/nvidia_modelopt-0.21.1-cp39-cp39-manylinux_2_28_aarch64.whl", hash = "sha256:6ba34f712a0c4333e3a37b4454d2f6e24d4dd45a99b5d7640a516f3d12ded8c8", size = 2590413 }, ] [package.optional-dependencies] @@ -664,6 +1246,7 @@ hf = [ { name = "accelerate", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, { name = "diffusers", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, { name = "huggingface-hub", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "peft", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, { name = "transformers", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, ] torch = [ @@ -699,6 +1282,85 @@ wheels = [ { url = "https://download.pytorch.org/whl/nightly/cu126/nvidia_nvtx_cu12-12.6.77-py3-none-manylinux2014_x86_64.whl" }, ] +[[package]] +name = "onnx" +version = "1.17.0" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "numpy", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "protobuf", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/9a/54/0e385c26bf230d223810a9c7d06628d954008a5e5e4b73ee26ef02327282/onnx-1.17.0.tar.gz", hash = "sha256:48ca1a91ff73c1d5e3ea2eef20ae5d0e709bb8a2355ed798ffc2169753013fd3", size = 12165120 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/75/0d/831807a18db2a5e8f7813848c59272b904a4ef3939fe4d1288cbce9ea735/onnx-1.17.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d545335cb49d4d8c47cc803d3a805deb7ad5d9094dc67657d66e568610a36d7d", size = 15908420 }, + { url = "https://files.pythonhosted.org/packages/dd/5b/c4f95dbe652d14aeba9afaceb177e9ffc48ac3c03048dd3f872f26f07e34/onnx-1.17.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3193a3672fc60f1a18c0f4c93ac81b761bc72fd8a6c2035fa79ff5969f07713e", size = 16046244 }, + { url = "https://files.pythonhosted.org/packages/7b/e3/cc80110e5996ca61878f7b4c73c7a286cd88918ff35eacb60dc75ab11ef5/onnx-1.17.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f01a4b63d4e1d8ec3e2f069e7b798b2955810aa434f7361f01bc8ca08d69cce4", size = 15908949 }, + { url = "https://files.pythonhosted.org/packages/b1/2f/91092557ed478e323a2b4471e2081fdf88d1dd52ae988ceaf7db4e4506ff/onnx-1.17.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4a183c6178be001bf398260e5ac2c927dc43e7746e8638d6c05c20e321f8c949", size = 16048190 }, + { url = "https://files.pythonhosted.org/packages/f0/6c/f040652277f514ecd81b7251841f96caa5538365af7df07f86c6018cda2b/onnx-1.17.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3d955ba2939878a520a97614bcf2e79c1df71b29203e8ced478fa78c9a9c63c2", size = 15907522 }, + { url = "https://files.pythonhosted.org/packages/3d/7c/67f4952d1b56b3f74a154b97d0dd0630d525923b354db117d04823b8b49b/onnx-1.17.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4f3fb5cc4e2898ac5312a7dc03a65133dd2abf9a5e520e69afb880a7251ec97a", size = 16046307 }, + { url = "https://files.pythonhosted.org/packages/61/94/d753c230d56234dd01ad939590a2ed33221b57c61abe513ff6823a69af6e/onnx-1.17.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3e19fd064b297f7773b4c1150f9ce6213e6d7d041d7a9201c0d348041009cdcd", size = 15908316 }, + { url = "https://files.pythonhosted.org/packages/3d/da/c19d0f20d310045f4701d75ecba4f765153251d48a32f27a5d6b0a7e3799/onnx-1.17.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8167295f576055158a966161f8ef327cb491c06ede96cc23392be6022071b6ed", size = 16046488 }, +] + +[[package]] +name = "onnx-graphsurgeon" +version = "0.5.2" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "numpy", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "onnx", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, +] +wheels = [ + { url = "https://files.pythonhosted.org/packages/0d/20/93e7143af3a0b3b3d9f3306bfc46e55d0d307242b4c1bf36ff108460e5a3/onnx_graphsurgeon-0.5.2-py2.py3-none-any.whl", hash = "sha256:10c130d6129fdeee02945f8103b5b112e6fd4d9b356e2dd3e80f53e0ebee7b5c", size = 56430 }, +] + +[[package]] +name = "openai" +version = "1.54.3" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "anyio", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "distro", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "httpx", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "jiter", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "pydantic", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "sniffio", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "tqdm", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "typing-extensions", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/4b/d8/ff02ef5255bf1b0c75f6badced772b4fd5c29a517d2a0be6f4e0a65a3227/openai-1.54.3.tar.gz", hash = "sha256:7511b74eeb894ac0b0253dc71f087a15d2e4d71d22d0088767205143d880cca6", size = 313963 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/77/85/e7adeee84edd24c6cc119b2ccaaacd9579c6a2c7f72d05e936ea6b33594e/openai-1.54.3-py3-none-any.whl", hash = "sha256:f18dbaf09c50d70c4185b892a2a553f80681d1d866323a2da7f7be2f688615d5", size = 389619 }, +] + +[[package]] +name = "optimum" +version = "1.23.3" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "coloredlogs", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "datasets", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "huggingface-hub", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "numpy", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "packaging", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "sympy", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "torch", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "transformers", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/f5/95/44eb569e2a70f9c63dd75f80fea8495eec464c29b988188ebcae940a6470/optimum-1.23.3.tar.gz", hash = "sha256:2089bd73d1232686473a80effd53800f8a8c385c02126e80d35c07227c1b9bf5", size = 341546 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/48/33/97cf226c47e4cf5a79159668732038cdd6c0199c72782d5b5a0db54f9a2d/optimum-1.23.3-py3-none-any.whl", hash = "sha256:ac34b497310e74e919e8eb3bc01cfea48bca304ade3e3ce8a7707d125120001a", size = 424070 }, +] + +[[package]] +name = "ordered-set" +version = "4.1.0" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/4c/ca/bfac8bc689799bcca4157e0e0ced07e70ce125193fc2e166d2e685b7e2fe/ordered-set-4.1.0.tar.gz", hash = "sha256:694a8e44c87657c59292ede72891eb91d34131f6531463aab3009191c77364a8", size = 12826 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/33/55/af02708f230eb77084a299d7b08175cff006dea4f2721074b92cdb0296c0/ordered_set-4.1.0-py3-none-any.whl", hash = "sha256:046e1132c71fcf3330438a539928932caf51ddbc582496833e23de611de14562", size = 7634 }, +] + [[package]] name = "packaging" version = "24.1" @@ -707,6 +1369,31 @@ wheels = [ { url = "https://download.pytorch.org/whl/nightly/packaging-24.1-py3-none-any.whl" }, ] +[[package]] +name = "pandas" +version = "2.2.3" +source = { registry = "https://download.pytorch.org/whl/nightly/cu126" } +dependencies = [ + { name = "numpy", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "python-dateutil", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "pytz", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "tzdata", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, +] +wheels = [ + { url = "https://download.pytorch.org/whl/nightly/pandas-2.2.3-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/pandas-2.2.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/pandas-2.2.3-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/pandas-2.2.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/pandas-2.2.3-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/pandas-2.2.3-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/pandas-2.2.3-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/pandas-2.2.3-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/pandas-2.2.3-cp313-cp313t-manylinux2014_aarch64.manylinux_2_17_aarch64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/pandas-2.2.3-cp313-cp313t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/pandas-2.2.3-cp39-cp39-manylinux2014_aarch64.manylinux_2_17_aarch64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/pandas-2.2.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, +] + [[package]] name = "parameterized" version = "0.9.0" @@ -734,6 +1421,27 @@ wheels = [ { url = "https://files.pythonhosted.org/packages/cc/20/ff623b09d963f88bfde16306a54e12ee5ea43e9b597108672ff3a408aad6/pathspec-0.12.1-py3-none-any.whl", hash = "sha256:a0d503e138a4c123b27490a4f7beda6a01c6f288df0e4a8b79c7eb0dc7b4cc08", size = 31191 }, ] +[[package]] +name = "peft" +version = "0.14.0" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "accelerate", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "huggingface-hub", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "numpy", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "packaging", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "psutil", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "pyyaml", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "safetensors", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "torch", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "tqdm", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "transformers", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/21/33/fb0c31eaa8162c01e9250b21aa65d46a5339f17a818a97c68391db2ff44b/peft-0.14.0.tar.gz", hash = "sha256:546d69af7b42f5ef715a3d3261ed818bc917ae6055e5d7e187ed3f2c76ad72dc", size = 411902 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/88/05/e58e3aaa36544d30a917814e336fc65a746f708e5874945e92999bc22fa3/peft-0.14.0-py3-none-any.whl", hash = "sha256:2f04f3a870c3baf30f15e7dcaa5dd70d3e54cfdd146d3c6c187735d3ae0a0700", size = 374831 }, +] + [[package]] name = "pexpect" version = "4.9.0" @@ -748,31 +1456,42 @@ wheels = [ [[package]] name = "pillow" -version = "11.0.0" -source = { registry = "https://download.pytorch.org/whl/nightly/cu126" } -wheels = [ - { url = "https://download.pytorch.org/whl/nightly/pillow-11.0.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pillow-11.0.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pillow-11.0.0-cp310-cp310-manylinux_2_28_aarch64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pillow-11.0.0-cp310-cp310-manylinux_2_28_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pillow-11.0.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pillow-11.0.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pillow-11.0.0-cp311-cp311-manylinux_2_28_aarch64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pillow-11.0.0-cp311-cp311-manylinux_2_28_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pillow-11.0.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pillow-11.0.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pillow-11.0.0-cp312-cp312-manylinux_2_28_aarch64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pillow-11.0.0-cp312-cp312-manylinux_2_28_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pillow-11.0.0-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pillow-11.0.0-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pillow-11.0.0-cp313-cp313-manylinux_2_28_aarch64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pillow-11.0.0-cp313-cp313-manylinux_2_28_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pillow-11.0.0-cp313-cp313t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pillow-11.0.0-cp313-cp313t-manylinux_2_28_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pillow-11.0.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pillow-11.0.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pillow-11.0.0-cp39-cp39-manylinux_2_28_aarch64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pillow-11.0.0-cp39-cp39-manylinux_2_28_x86_64.whl" }, +version = "10.3.0" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/ef/43/c50c17c5f7d438e836c169e343695534c38c77f60e7c90389bd77981bc21/pillow-10.3.0.tar.gz", hash = "sha256:9d2455fbf44c914840c793e89aa82d0e1763a14253a000743719ae5946814b2d", size = 46572854 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/bb/a5/7958a4c0941b611a7706db510b9a85939346990df55ea05ecdfffb2b050c/pillow-10.3.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:78618cdbccaa74d3f88d0ad6cb8ac3007f1a6fa5c6f19af64b55ca170bfa1edf", size = 4309181 }, + { url = "https://files.pythonhosted.org/packages/01/d7/0d3021e6c2da8f2a5d6f7e97ebf0bf540e69ebe3d0384c207401bfe88ef5/pillow-10.3.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:261ddb7ca91fcf71757979534fb4c128448b5b4c55cb6152d280312062f69599", size = 4420421 }, + { url = "https://files.pythonhosted.org/packages/88/3c/708d0fc162f3c7099254b488b80ec4aba2a7fbdb958c03279390cf6e1140/pillow-10.3.0-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:ce49c67f4ea0609933d01c0731b34b8695a7a748d6c8d186f95e7d085d2fe475", size = 4333092 }, + { url = "https://files.pythonhosted.org/packages/b5/a2/7a09695dc636bf8d0a1b63022f58701177b7dc6fad30f6d6bc343e5473a4/pillow-10.3.0-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:b14f16f94cbc61215115b9b1236f9c18403c15dd3c52cf629072afa9d54c1cbf", size = 4499372 }, + { url = "https://files.pythonhosted.org/packages/dd/b8/ff0e2a7f4bba4d0121bfcd06387ea28660d7497ea038f99640bb10015125/pillow-10.3.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:d33891be6df59d93df4d846640f0e46f1a807339f09e79a8040bc887bdcd7ed3", size = 4528038 }, + { url = "https://files.pythonhosted.org/packages/d5/9f/f19b94322353ca97e3b653255bf26b385ded07582f33eb6cd17f44d2b2bc/pillow-10.3.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:b50811d664d392f02f7761621303eba9d1b056fb1868c8cdf4231279645c25f5", size = 4592192 }, + { url = "https://files.pythonhosted.org/packages/1d/29/abda81a079cccd1840b0b7b13ad67ffac87cc66395ae20973027280e9f9f/pillow-10.3.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fdcbb4068117dfd9ce0138d068ac512843c52295ed996ae6dd1faf537b6dbc27", size = 4317733 }, + { url = "https://files.pythonhosted.org/packages/77/cd/5205fb43a6000d424291b0525b8201004700d9a34e034517ac4dfdc6eed5/pillow-10.3.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9797a6c8fe16f25749b371c02e2ade0efb51155e767a971c61734b1bf6293994", size = 4429430 }, + { url = "https://files.pythonhosted.org/packages/8c/bb/9e8d2b1b54235bd44139ee387beeb65ad9d8d755b5c01f817070c6dabea7/pillow-10.3.0-cp311-cp311-manylinux_2_28_aarch64.whl", hash = "sha256:9e91179a242bbc99be65e139e30690e081fe6cb91a8e77faf4c409653de39451", size = 4341711 }, + { url = "https://files.pythonhosted.org/packages/81/ff/ad3c942d865f9e45ce84eeb31795e6d4d94e1f1eea51026d5154028510d7/pillow-10.3.0-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:1b87bd9d81d179bd8ab871603bd80d8645729939f90b71e62914e816a76fc6bd", size = 4507469 }, + { url = "https://files.pythonhosted.org/packages/ab/ab/30cd50a12d9afa2c412efcb8b37dd3f5f1da4bc77b984ddfbc776d96cf5b/pillow-10.3.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:81d09caa7b27ef4e61cb7d8fbf1714f5aec1c6b6c5270ee53504981e6e9121ad", size = 4533491 }, + { url = "https://files.pythonhosted.org/packages/1f/f0/07419615ffa852cded35dfa3337bf70788f232a3dfe622b97d5eb0c32674/pillow-10.3.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:048ad577748b9fa4a99a0548c64f2cb8d672d5bf2e643a739ac8faff1164238c", size = 4598334 }, + { url = "https://files.pythonhosted.org/packages/53/7b/4f7b153a776725a87797d744ea1c73b83ac0b723f5e379297605dee118eb/pillow-10.3.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9d3bea1c75f8c53ee4d505c3e67d8c158ad4df0d83170605b50b64025917f338", size = 4321427 }, + { url = "https://files.pythonhosted.org/packages/45/08/d2cc751b790e77464f8648aa707e2327d6da5d95cf236a532e99c2e7a499/pillow-10.3.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:19aeb96d43902f0a783946a0a87dbdad5c84c936025b8419da0a0cd7724356b1", size = 4435915 }, + { url = "https://files.pythonhosted.org/packages/ef/97/f69d1932cf45bf5bd9fa1e2ae57bdf716524faa4fa9fb7dc62cdb1a19113/pillow-10.3.0-cp312-cp312-manylinux_2_28_aarch64.whl", hash = "sha256:74d28c17412d9caa1066f7a31df8403ec23d5268ba46cd0ad2c50fb82ae40462", size = 4347392 }, + { url = "https://files.pythonhosted.org/packages/c6/c1/3521ddb9c1f3ac106af3e4512a98c785b6ed8a39e0f778480b8a4d340165/pillow-10.3.0-cp312-cp312-manylinux_2_28_x86_64.whl", hash = "sha256:ff61bfd9253c3915e6d41c651d5f962da23eda633cf02262990094a18a55371a", size = 4514536 }, + { url = "https://files.pythonhosted.org/packages/c0/6f/347c241904a6514e59515284b01ba6f61765269a0d1a19fd2e6cbe331c8a/pillow-10.3.0-cp312-cp312-musllinux_1_1_aarch64.whl", hash = "sha256:d886f5d353333b4771d21267c7ecc75b710f1a73d72d03ca06df49b09015a9ef", size = 4555987 }, + { url = "https://files.pythonhosted.org/packages/c3/e2/3cc490c6b2e262713da82ce849c34bd8e6c31242afb53be8595d820b9877/pillow-10.3.0-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:4b5ec25d8b17217d635f8935dbc1b9aa5907962fae29dff220f2659487891cd3", size = 4623526 }, + { url = "https://files.pythonhosted.org/packages/ef/cb/a3c20e6fc07bead46aa548b97dd05854424938e0544c9f788008a8c0fb77/pillow-10.3.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4203efca580f0dd6f882ca211f923168548f7ba334c189e9eab1178ab840bf60", size = 4305324 }, + { url = "https://files.pythonhosted.org/packages/15/5c/2e16159554296a10017bfad367d495909f863abf7ea506f24fff8e6799b3/pillow-10.3.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3102045a10945173d38336f6e71a8dc71bcaeed55c3123ad4af82c52807b9375", size = 4416226 }, + { url = "https://files.pythonhosted.org/packages/6e/44/53244b128f0edc837bfa07706874eb02423929150647feaa7e71a20dd694/pillow-10.3.0-cp39-cp39-manylinux_2_28_aarch64.whl", hash = "sha256:6fb1b30043271ec92dc65f6d9f0b7a830c210b8a96423074b15c7bc999975f57", size = 4329482 }, + { url = "https://files.pythonhosted.org/packages/f5/6d/52e82352670e850f468de9e6bccced4202a09f58e7ea5ecdbf08283d85cb/pillow-10.3.0-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:1dfc94946bc60ea375cc39cff0b8da6c7e5f8fcdc1d946beb8da5c216156ddd8", size = 4496414 }, + { url = "https://files.pythonhosted.org/packages/c6/5d/c21156798e72362e2ef7b6a9d034a1f3a542f7cd3cbf5bedd10a71fd32c2/pillow-10.3.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:b09b86b27a064c9624d0a6c54da01c1beaf5b6cadfa609cf63789b1d08a797b9", size = 4522768 }, + { url = "https://files.pythonhosted.org/packages/64/bc/e9e7b4417eebb8231bc4bd62a56609a47f153f26430c17bd3c4d4ecf9e90/pillow-10.3.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:d3b2348a78bc939b4fed6552abfd2e7988e0f81443ef3911a4b8498ca084f6eb", size = 4588088 }, + { url = "https://files.pythonhosted.org/packages/73/9f/cf2523a1c3a98afd0052b11d12d866453a60151bfc5876620e88cd5be55c/pillow-10.3.0-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:048eeade4c33fdf7e08da40ef402e748df113fd0b4584e32c4af74fe78baaeb2", size = 3414179 }, + { url = "https://files.pythonhosted.org/packages/12/d1/010dca4eaaaeb9da9edb702d2f663b6dac98ff5e84ce09e9d82f96c6a9f3/pillow-10.3.0-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9e2ec1e921fd07c7cda7962bad283acc2f2a9ccc1b971ee4b216b75fad6f0463", size = 3468521 }, + { url = "https://files.pythonhosted.org/packages/ff/4c/8c7e9830ccca3219cdf4c1bdd3b0664025c91034a29242aedec5a997cbfe/pillow-10.3.0-pp310-pypy310_pp73-manylinux_2_28_aarch64.whl", hash = "sha256:4c8e73e99da7db1b4cad7f8d682cf6abad7844da39834c288fbfa394a47bbced", size = 3455893 }, + { url = "https://files.pythonhosted.org/packages/aa/e3/a84acfed7c3ccb23ff58fa68ae9f3ec071d63cfb7885edb6eb48bbc907f7/pillow-10.3.0-pp310-pypy310_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:16563993329b79513f59142a6b02055e10514c1a8e86dca8b48a893e33cf91e3", size = 3557538 }, + { url = "https://files.pythonhosted.org/packages/78/80/8f1028ff93edb59d57cca0b3b7687ee5190f420b580d25fa96991958f400/pillow-10.3.0-pp39-pypy39_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:793b4e24db2e8742ca6423d3fde8396db336698c55cd34b660663ee9e45ed37f", size = 3414098 }, + { url = "https://files.pythonhosted.org/packages/a9/2e/aff844131bf2987d670daebf9b00e4f964f5a2de51b88b82e7c0bcaa13a0/pillow-10.3.0-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d93480005693d247f8346bc8ee28c72a2191bdf1f6b5db469c096c0c867ac015", size = 3468295 }, + { url = "https://files.pythonhosted.org/packages/35/32/186ec6365fca279e4d70e9fb43f5adea013ec2f020ca03ec2966648573f9/pillow-10.3.0-pp39-pypy39_pp73-manylinux_2_28_aarch64.whl", hash = "sha256:c83341b89884e2b2e55886e8fbbf37c3fa5efd6c8907124aeb72f285ae5696e5", size = 3455842 }, + { url = "https://files.pythonhosted.org/packages/36/8d/e312d570c7775576c65c5fc30ca22e2de348e7d7f00566b1087bd4947aef/pillow-10.3.0-pp39-pypy39_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:1a1d1915db1a4fdb2754b9de292642a39a7fb28f1736699527bb649484fb966a", size = 3557549 }, ] [[package]] @@ -793,6 +1512,14 @@ wheels = [ { url = "https://files.pythonhosted.org/packages/88/5f/e351af9a41f866ac3f1fac4ca0613908d9a41741cfcf2228f4ad853b697d/pluggy-1.5.0-py3-none-any.whl", hash = "sha256:44e1ad92c8ca002de6377e165f3e0f1be63266ab4d554740532335b9d75ea669", size = 20556 }, ] +[[package]] +name = "polygraphy" +version = "0.49.9" +source = { registry = "https://pypi.org/simple" } +wheels = [ + { url = "https://files.pythonhosted.org/packages/4a/f5/a2b20c677c1a856cc9e08cd0b5a5105450ed5253e369e938ddd31d91c547/polygraphy-0.49.9-py2.py3-none-any.whl", hash = "sha256:62ae22825efdd3288222e5b1d2d791fe58e87844fcd848bcd1251fbce02ba956", size = 346910 }, +] + [[package]] name = "pre-commit" version = "4.0.1" @@ -821,6 +1548,81 @@ wheels = [ { url = "https://files.pythonhosted.org/packages/a9/6a/fd08d94654f7e67c52ca30523a178b3f8ccc4237fce4be90d39c938a831a/prompt_toolkit-3.0.48-py3-none-any.whl", hash = "sha256:f49a827f90062e411f1ce1f854f2aedb3c23353244f8108b89283587397ac10e", size = 386595 }, ] +[[package]] +name = "propcache" +version = "0.2.1" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/20/c8/2a13f78d82211490855b2fb303b6721348d0787fdd9a12ac46d99d3acde1/propcache-0.2.1.tar.gz", hash = "sha256:3f77ce728b19cb537714499928fe800c3dda29e8d9428778fc7c186da4c09a64", size = 41735 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/4d/3d/31c9c29ee7192defc05aa4d01624fd85a41cf98e5922aaed206017329944/propcache-0.2.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f9479aa06a793c5aeba49ce5c5692ffb51fcd9a7016e017d555d5e2b0045d212", size = 204809 }, + { url = "https://files.pythonhosted.org/packages/10/a1/e4050776f4797fc86140ac9a480d5dc069fbfa9d499fe5c5d2fa1ae71f07/propcache-0.2.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d9631c5e8b5b3a0fda99cb0d29c18133bca1e18aea9effe55adb3da1adef80d3", size = 219109 }, + { url = "https://files.pythonhosted.org/packages/c9/c0/e7ae0df76343d5e107d81e59acc085cea5fd36a48aa53ef09add7503e888/propcache-0.2.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:3156628250f46a0895f1f36e1d4fbe062a1af8718ec3ebeb746f1d23f0c5dc4d", size = 217368 }, + { url = "https://files.pythonhosted.org/packages/fc/e1/e0a2ed6394b5772508868a977d3238f4afb2eebaf9976f0b44a8d347ad63/propcache-0.2.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6b6fb63ae352e13748289f04f37868099e69dba4c2b3e271c46061e82c745634", size = 205124 }, + { url = "https://files.pythonhosted.org/packages/50/c1/e388c232d15ca10f233c778bbdc1034ba53ede14c207a72008de45b2db2e/propcache-0.2.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:887d9b0a65404929641a9fabb6452b07fe4572b269d901d622d8a34a4e9043b2", size = 195463 }, + { url = "https://files.pythonhosted.org/packages/0a/fd/71b349b9def426cc73813dbd0f33e266de77305e337c8c12bfb0a2a82bfb/propcache-0.2.1-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:a96dc1fa45bd8c407a0af03b2d5218392729e1822b0c32e62c5bf7eeb5fb3958", size = 198358 }, + { url = "https://files.pythonhosted.org/packages/02/f2/d7c497cd148ebfc5b0ae32808e6c1af5922215fe38c7a06e4e722fe937c8/propcache-0.2.1-cp310-cp310-musllinux_1_2_armv7l.whl", hash = "sha256:a7e65eb5c003a303b94aa2c3852ef130230ec79e349632d030e9571b87c4698c", size = 195560 }, + { url = "https://files.pythonhosted.org/packages/bb/57/f37041bbe5e0dfed80a3f6be2612a3a75b9cfe2652abf2c99bef3455bbad/propcache-0.2.1-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:999779addc413181912e984b942fbcc951be1f5b3663cd80b2687758f434c583", size = 196895 }, + { url = "https://files.pythonhosted.org/packages/83/36/ae3cc3e4f310bff2f064e3d2ed5558935cc7778d6f827dce74dcfa125304/propcache-0.2.1-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:19a0f89a7bb9d8048d9c4370c9c543c396e894c76be5525f5e1ad287f1750ddf", size = 207124 }, + { url = "https://files.pythonhosted.org/packages/8c/c4/811b9f311f10ce9d31a32ff14ce58500458443627e4df4ae9c264defba7f/propcache-0.2.1-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:1ac2f5fe02fa75f56e1ad473f1175e11f475606ec9bd0be2e78e4734ad575034", size = 210442 }, + { url = "https://files.pythonhosted.org/packages/18/dd/a1670d483a61ecac0d7fc4305d91caaac7a8fc1b200ea3965a01cf03bced/propcache-0.2.1-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:574faa3b79e8ebac7cb1d7930f51184ba1ccf69adfdec53a12f319a06030a68b", size = 203219 }, + { url = "https://files.pythonhosted.org/packages/03/7a/793aa12f0537b2e520bf09f4c6833706b63170a211ad042ca71cbf79d9cb/propcache-0.2.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b480c6a4e1138e1aa137c0079b9b6305ec6dcc1098a8ca5196283e8a49df95a9", size = 232136 }, + { url = "https://files.pythonhosted.org/packages/f1/38/b921b3168d72111769f648314100558c2ea1d52eb3d1ba7ea5c4aa6f9848/propcache-0.2.1-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d27b84d5880f6d8aa9ae3edb253c59d9f6642ffbb2c889b78b60361eed449787", size = 239706 }, + { url = "https://files.pythonhosted.org/packages/14/29/4636f500c69b5edea7786db3c34eb6166f3384b905665ce312a6e42c720c/propcache-0.2.1-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:857112b22acd417c40fa4595db2fe28ab900c8c5fe4670c7989b1c0230955465", size = 238531 }, + { url = "https://files.pythonhosted.org/packages/85/14/01fe53580a8e1734ebb704a3482b7829a0ef4ea68d356141cf0994d9659b/propcache-0.2.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:cf6c4150f8c0e32d241436526f3c3f9cbd34429492abddbada2ffcff506c51af", size = 231063 }, + { url = "https://files.pythonhosted.org/packages/33/5c/1d961299f3c3b8438301ccfbff0143b69afcc30c05fa28673cface692305/propcache-0.2.1-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:66d4cfda1d8ed687daa4bc0274fcfd5267873db9a5bc0418c2da19273040eeb7", size = 220134 }, + { url = "https://files.pythonhosted.org/packages/00/d0/ed735e76db279ba67a7d3b45ba4c654e7b02bc2f8050671ec365d8665e21/propcache-0.2.1-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:c2f992c07c0fca81655066705beae35fc95a2fa7366467366db627d9f2ee097f", size = 220009 }, + { url = "https://files.pythonhosted.org/packages/75/90/ee8fab7304ad6533872fee982cfff5a53b63d095d78140827d93de22e2d4/propcache-0.2.1-cp311-cp311-musllinux_1_2_armv7l.whl", hash = "sha256:4a571d97dbe66ef38e472703067021b1467025ec85707d57e78711c085984e54", size = 212199 }, + { url = "https://files.pythonhosted.org/packages/eb/ec/977ffaf1664f82e90737275873461695d4c9407d52abc2f3c3e24716da13/propcache-0.2.1-cp311-cp311-musllinux_1_2_i686.whl", hash = "sha256:bb6178c241278d5fe853b3de743087be7f5f4c6f7d6d22a3b524d323eecec505", size = 214827 }, + { url = "https://files.pythonhosted.org/packages/57/48/031fb87ab6081764054821a71b71942161619549396224cbb242922525e8/propcache-0.2.1-cp311-cp311-musllinux_1_2_ppc64le.whl", hash = "sha256:ad1af54a62ffe39cf34db1aa6ed1a1873bd548f6401db39d8e7cd060b9211f82", size = 228009 }, + { url = "https://files.pythonhosted.org/packages/1a/06/ef1390f2524850838f2390421b23a8b298f6ce3396a7cc6d39dedd4047b0/propcache-0.2.1-cp311-cp311-musllinux_1_2_s390x.whl", hash = "sha256:e7048abd75fe40712005bcfc06bb44b9dfcd8e101dda2ecf2f5aa46115ad07ca", size = 231638 }, + { url = "https://files.pythonhosted.org/packages/38/2a/101e6386d5a93358395da1d41642b79c1ee0f3b12e31727932b069282b1d/propcache-0.2.1-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:160291c60081f23ee43d44b08a7e5fb76681221a8e10b3139618c5a9a291b84e", size = 222788 }, + { url = "https://files.pythonhosted.org/packages/7f/14/7ae06a6cf2a2f1cb382586d5a99efe66b0b3d0c6f9ac2f759e6f7af9d7cf/propcache-0.2.1-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:049324ee97bb67285b49632132db351b41e77833678432be52bdd0289c0e05e4", size = 241869 }, + { url = "https://files.pythonhosted.org/packages/cc/59/227a78be960b54a41124e639e2c39e8807ac0c751c735a900e21315f8c2b/propcache-0.2.1-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:1cd9a1d071158de1cc1c71a26014dcdfa7dd3d5f4f88c298c7f90ad6f27bb46d", size = 247884 }, + { url = "https://files.pythonhosted.org/packages/84/58/f62b4ffaedf88dc1b17f04d57d8536601e4e030feb26617228ef930c3279/propcache-0.2.1-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:98110aa363f1bb4c073e8dcfaefd3a5cea0f0834c2aab23dda657e4dab2f53b5", size = 248486 }, + { url = "https://files.pythonhosted.org/packages/1c/07/ebe102777a830bca91bbb93e3479cd34c2ca5d0361b83be9dbd93104865e/propcache-0.2.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:647894f5ae99c4cf6bb82a1bb3a796f6e06af3caa3d32e26d2350d0e3e3faf24", size = 243649 }, + { url = "https://files.pythonhosted.org/packages/ed/bc/4f7aba7f08f520376c4bb6a20b9a981a581b7f2e385fa0ec9f789bb2d362/propcache-0.2.1-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:bfd3223c15bebe26518d58ccf9a39b93948d3dcb3e57a20480dfdd315356baff", size = 229103 }, + { url = "https://files.pythonhosted.org/packages/fe/d5/04ac9cd4e51a57a96f78795e03c5a0ddb8f23ec098b86f92de028d7f2a6b/propcache-0.2.1-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:d71264a80f3fcf512eb4f18f59423fe82d6e346ee97b90625f283df56aee103f", size = 226607 }, + { url = "https://files.pythonhosted.org/packages/e3/f0/24060d959ea41d7a7cc7fdbf68b31852331aabda914a0c63bdb0e22e96d6/propcache-0.2.1-cp312-cp312-musllinux_1_2_armv7l.whl", hash = "sha256:e73091191e4280403bde6c9a52a6999d69cdfde498f1fdf629105247599b57ec", size = 221153 }, + { url = "https://files.pythonhosted.org/packages/77/a7/3ac76045a077b3e4de4859a0753010765e45749bdf53bd02bc4d372da1a0/propcache-0.2.1-cp312-cp312-musllinux_1_2_i686.whl", hash = "sha256:3935bfa5fede35fb202c4b569bb9c042f337ca4ff7bd540a0aa5e37131659348", size = 222151 }, + { url = "https://files.pythonhosted.org/packages/e7/af/5e29da6f80cebab3f5a4dcd2a3240e7f56f2c4abf51cbfcc99be34e17f0b/propcache-0.2.1-cp312-cp312-musllinux_1_2_ppc64le.whl", hash = "sha256:f508b0491767bb1f2b87fdfacaba5f7eddc2f867740ec69ece6d1946d29029a6", size = 233812 }, + { url = "https://files.pythonhosted.org/packages/8c/89/ebe3ad52642cc5509eaa453e9f4b94b374d81bae3265c59d5c2d98efa1b4/propcache-0.2.1-cp312-cp312-musllinux_1_2_s390x.whl", hash = "sha256:1672137af7c46662a1c2be1e8dc78cb6d224319aaa40271c9257d886be4363a6", size = 238829 }, + { url = "https://files.pythonhosted.org/packages/e9/2f/6b32f273fa02e978b7577159eae7471b3cfb88b48563b1c2578b2d7ca0bb/propcache-0.2.1-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:b74c261802d3d2b85c9df2dfb2fa81b6f90deeef63c2db9f0e029a3cac50b518", size = 230704 }, + { url = "https://files.pythonhosted.org/packages/4e/3e/021b6cd86c0acc90d74784ccbb66808b0bd36067a1bf3e2deb0f3845f618/propcache-0.2.1-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ba278acf14471d36316159c94a802933d10b6a1e117b8554fe0d0d9b75c9d536", size = 224819 }, + { url = "https://files.pythonhosted.org/packages/3c/57/c2fdeed1b3b8918b1770a133ba5c43ad3d78e18285b0c06364861ef5cc38/propcache-0.2.1-cp313-cp313-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4e6281aedfca15301c41f74d7005e6e3f4ca143584ba696ac69df4f02f40d629", size = 229625 }, + { url = "https://files.pythonhosted.org/packages/9d/81/70d4ff57bf2877b5780b466471bebf5892f851a7e2ca0ae7ffd728220281/propcache-0.2.1-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:5b750a8e5a1262434fb1517ddf64b5de58327f1adc3524a5e44c2ca43305eb0b", size = 232934 }, + { url = "https://files.pythonhosted.org/packages/3c/b9/bb51ea95d73b3fb4100cb95adbd4e1acaf2cbb1fd1083f5468eeb4a099a8/propcache-0.2.1-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bf72af5e0fb40e9babf594308911436c8efde3cb5e75b6f206c34ad18be5c052", size = 227361 }, + { url = "https://files.pythonhosted.org/packages/f1/20/3c6d696cd6fd70b29445960cc803b1851a1131e7a2e4ee261ee48e002bcd/propcache-0.2.1-cp313-cp313-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b2d0a12018b04f4cb820781ec0dffb5f7c7c1d2a5cd22bff7fb055a2cb19ebce", size = 213904 }, + { url = "https://files.pythonhosted.org/packages/a1/cb/1593bfc5ac6d40c010fa823f128056d6bc25b667f5393781e37d62f12005/propcache-0.2.1-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:e800776a79a5aabdb17dcc2346a7d66d0777e942e4cd251defeb084762ecd17d", size = 212632 }, + { url = "https://files.pythonhosted.org/packages/6d/5c/e95617e222be14a34c709442a0ec179f3207f8a2b900273720501a70ec5e/propcache-0.2.1-cp313-cp313-musllinux_1_2_armv7l.whl", hash = "sha256:4160d9283bd382fa6c0c2b5e017acc95bc183570cd70968b9202ad6d8fc48dce", size = 207897 }, + { url = "https://files.pythonhosted.org/packages/8e/3b/56c5ab3dc00f6375fbcdeefdede5adf9bee94f1fab04adc8db118f0f9e25/propcache-0.2.1-cp313-cp313-musllinux_1_2_i686.whl", hash = "sha256:30b43e74f1359353341a7adb783c8f1b1c676367b011709f466f42fda2045e95", size = 208118 }, + { url = "https://files.pythonhosted.org/packages/86/25/d7ef738323fbc6ebcbce33eb2a19c5e07a89a3df2fded206065bd5e868a9/propcache-0.2.1-cp313-cp313-musllinux_1_2_ppc64le.whl", hash = "sha256:58791550b27d5488b1bb52bc96328456095d96206a250d28d874fafe11b3dfaf", size = 217851 }, + { url = "https://files.pythonhosted.org/packages/b3/77/763e6cef1852cf1ba740590364ec50309b89d1c818e3256d3929eb92fabf/propcache-0.2.1-cp313-cp313-musllinux_1_2_s390x.whl", hash = "sha256:0f022d381747f0dfe27e99d928e31bc51a18b65bb9e481ae0af1380a6725dd1f", size = 222630 }, + { url = "https://files.pythonhosted.org/packages/4f/e9/0f86be33602089c701696fbed8d8c4c07b6ee9605c5b7536fd27ed540c5b/propcache-0.2.1-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:297878dc9d0a334358f9b608b56d02e72899f3b8499fc6044133f0d319e2ec30", size = 216269 }, + { url = "https://files.pythonhosted.org/packages/22/59/6fe80a3fe7720f715f2c0f6df250dacbd7cad42832410dbd84c719c52f78/propcache-0.2.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5eee736daafa7af6d0a2dc15cc75e05c64f37fc37bafef2e00d77c14171c2097", size = 207792 }, + { url = "https://files.pythonhosted.org/packages/4a/68/584cd51dd8f4d0f5fff5b128ce0cdb257cde903898eecfb92156bbc2c780/propcache-0.2.1-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:f7a31fc1e1bd362874863fdeed71aed92d348f5336fd84f2197ba40c59f061bd", size = 223280 }, + { url = "https://files.pythonhosted.org/packages/85/cb/4c3528460c41e61b06ec3f970c0f89f87fa21f63acac8642ed81a886c164/propcache-0.2.1-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:cba4cfa1052819d16699e1d55d18c92b6e094d4517c41dd231a8b9f87b6fa681", size = 221293 }, + { url = "https://files.pythonhosted.org/packages/69/c0/560e050aa6d31eeece3490d1174da508f05ab27536dfc8474af88b97160a/propcache-0.2.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f089118d584e859c62b3da0892b88a83d611c2033ac410e929cb6754eec0ed16", size = 208259 }, + { url = "https://files.pythonhosted.org/packages/0c/87/d6c86a77632eb1ba86a328e3313159f246e7564cb5951e05ed77555826a0/propcache-0.2.1-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:781e65134efaf88feb447e8c97a51772aa75e48b794352f94cb7ea717dedda0d", size = 198632 }, + { url = "https://files.pythonhosted.org/packages/3a/2b/3690ea7b662dc762ab7af5f3ef0e2d7513c823d193d7b2a1b4cda472c2be/propcache-0.2.1-cp39-cp39-musllinux_1_2_aarch64.whl", hash = "sha256:31f5af773530fd3c658b32b6bdc2d0838543de70eb9a2156c03e410f7b0d3aae", size = 203516 }, + { url = "https://files.pythonhosted.org/packages/4d/b5/afe716c16c23c77657185c257a41918b83e03993b6ccdfa748e5e7d328e9/propcache-0.2.1-cp39-cp39-musllinux_1_2_armv7l.whl", hash = "sha256:a7a078f5d37bee6690959c813977da5291b24286e7b962e62a94cec31aa5188b", size = 199402 }, + { url = "https://files.pythonhosted.org/packages/a4/c0/2d2df3aa7f8660d0d4cc4f1e00490c48d5958da57082e70dea7af366f876/propcache-0.2.1-cp39-cp39-musllinux_1_2_i686.whl", hash = "sha256:cea7daf9fc7ae6687cf1e2c049752f19f146fdc37c2cc376e7d0032cf4f25347", size = 200528 }, + { url = "https://files.pythonhosted.org/packages/21/c8/65ac9142f5e40c8497f7176e71d18826b09e06dd4eb401c9a4ee41aa9c74/propcache-0.2.1-cp39-cp39-musllinux_1_2_ppc64le.whl", hash = "sha256:8b3489ff1ed1e8315674d0775dc7d2195fb13ca17b3808721b54dbe9fd020faf", size = 211254 }, + { url = "https://files.pythonhosted.org/packages/09/e4/edb70b447a1d8142df51ec7511e84aa64d7f6ce0a0fdf5eb55363cdd0935/propcache-0.2.1-cp39-cp39-musllinux_1_2_s390x.whl", hash = "sha256:9403db39be1393618dd80c746cb22ccda168efce239c73af13c3763ef56ffc04", size = 214589 }, + { url = "https://files.pythonhosted.org/packages/cb/02/817f309ec8d8883287781d6d9390f80b14db6e6de08bc659dfe798a825c2/propcache-0.2.1-cp39-cp39-musllinux_1_2_x86_64.whl", hash = "sha256:5d97151bc92d2b2578ff7ce779cdb9174337390a535953cbb9452fb65164c587", size = 207283 }, + { url = "https://files.pythonhosted.org/packages/41/b6/c5319caea262f4821995dca2107483b94a3345d4607ad797c76cb9c36bcc/propcache-0.2.1-py3-none-any.whl", hash = "sha256:52277518d6aae65536e9cea52d4e7fd2f7a66f4aa2d30ed3f2fcea620ace3c54", size = 11818 }, +] + +[[package]] +name = "protobuf" +version = "5.29.2" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/a5/73/4e6295c1420a9d20c9c351db3a36109b4c9aa601916cb7c6871e3196a1ca/protobuf-5.29.2.tar.gz", hash = "sha256:b2cc8e8bb7c9326996f0e160137b0861f1a82162502658df2951209d0cb0309e", size = 424901 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/e4/20/38fc33b60dcfb380507b99494aebe8c34b68b8ac7d32808c4cebda3f6f6b/protobuf-5.29.2-cp38-abi3-manylinux2014_aarch64.whl", hash = "sha256:494229ecd8c9009dd71eda5fd57528395d1eacdf307dbece6c12ad0dd09e912e", size = 319562 }, + { url = "https://files.pythonhosted.org/packages/90/4d/c3d61e698e0e41d926dbff6aa4e57428ab1a6fc3b5e1deaa6c9ec0fd45cf/protobuf-5.29.2-cp38-abi3-manylinux2014_x86_64.whl", hash = "sha256:b6b0d416bbbb9d4fbf9d0561dbfc4e324fd522f61f7af0fe0f282ab67b22477e", size = 319662 }, + { url = "https://files.pythonhosted.org/packages/f3/fd/c7924b4c2a1c61b8f4b64edd7a31ffacf63432135a2606f03a2f0d75a750/protobuf-5.29.2-py3-none-any.whl", hash = "sha256:fde4554c0e578a5a0bcc9a276339594848d1e89f9ea47b4427c80e5d72f90181", size = 172539 }, +] + [[package]] name = "psutil" version = "6.0.0" @@ -858,86 +1660,127 @@ wheels = [ { url = "https://files.pythonhosted.org/packages/8e/37/efad0257dc6e593a18957422533ff0f87ede7c9c6ea010a2177d738fb82f/pure_eval-0.2.3-py3-none-any.whl", hash = "sha256:1db8e35b67b3d218d818ae653e27f06c3aa420901fa7b081ca98cbedc874e0d0", size = 11842 }, ] +[[package]] +name = "pyarrow" +version = "18.1.0" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/7f/7b/640785a9062bb00314caa8a387abce547d2a420cf09bd6c715fe659ccffb/pyarrow-18.1.0.tar.gz", hash = "sha256:9386d3ca9c145b5539a1cfc75df07757dff870168c959b473a0bccbc3abc8c73", size = 1118671 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/a3/2a/526545a7464b5fb2fa6e2c4bad16ca90e59e1843025c534fd907b7f73e5a/pyarrow-18.1.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4f443122c8e31f4c9199cb23dca29ab9427cef990f283f80fe15b8e124bcc49b", size = 39213905 }, + { url = "https://files.pythonhosted.org/packages/8a/77/4b3fab91a30e19e233e738d0c5eca5a8f6dd05758bc349a2ca262c65de79/pyarrow-18.1.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c0a03da7f2758645d17b7b4f83c8bffeae5bbb7f974523fe901f36288d2eab71", size = 40128881 }, + { url = "https://files.pythonhosted.org/packages/aa/e2/a88e16c5e45e562449c52305bd3bc2f9d704295322d3434656e7ccac1444/pyarrow-18.1.0-cp310-cp310-manylinux_2_28_aarch64.whl", hash = "sha256:ba17845efe3aa358ec266cf9cc2800fa73038211fb27968bfa88acd09261a470", size = 38627517 }, + { url = "https://files.pythonhosted.org/packages/6d/84/8037c20005ccc7b869726465be0957bd9c29cfc88612962030f08292ad06/pyarrow-18.1.0-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:3c35813c11a059056a22a3bef520461310f2f7eea5c8a11ef9de7062a23f8d56", size = 40060187 }, + { url = "https://files.pythonhosted.org/packages/75/7e/332055ac913373e89256dce9d14b7708f55f7bd5be631456c897f0237738/pyarrow-18.1.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f39a2e0ed32a0970e4e46c262753417a60c43a3246972cfc2d3eb85aedd01b21", size = 39212135 }, + { url = "https://files.pythonhosted.org/packages/8c/64/5099cdb325828722ef7ffeba9a4696f238eb0cdeae227f831c2d77fcf1bd/pyarrow-18.1.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e31e9417ba9c42627574bdbfeada7217ad8a4cbbe45b9d6bdd4b62abbca4c6f6", size = 40125195 }, + { url = "https://files.pythonhosted.org/packages/83/88/1938d783727db1b178ff71bc6a6143d7939e406db83a9ec23cad3dad325c/pyarrow-18.1.0-cp311-cp311-manylinux_2_28_aarch64.whl", hash = "sha256:01c034b576ce0eef554f7c3d8c341714954be9b3f5d5bc7117006b85fcf302fe", size = 38641884 }, + { url = "https://files.pythonhosted.org/packages/5e/b5/9e14e9f7590e0eaa435ecea84dabb137284a4dbba7b3c337b58b65b76d95/pyarrow-18.1.0-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:f266a2c0fc31995a06ebd30bcfdb7f615d7278035ec5b1cd71c48d56daaf30b0", size = 40076877 }, + { url = "https://files.pythonhosted.org/packages/68/f9/29fb659b390312a7345aeb858a9d9c157552a8852522f2c8bad437c29c0a/pyarrow-18.1.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:543ad8459bc438efc46d29a759e1079436290bd583141384c6f7a1068ed6f992", size = 39203624 }, + { url = "https://files.pythonhosted.org/packages/6e/f6/19360dae44200e35753c5c2889dc478154cd78e61b1f738514c9f131734d/pyarrow-18.1.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0743e503c55be0fdb5c08e7d44853da27f19dc854531c0570f9f394ec9671d54", size = 40139341 }, + { url = "https://files.pythonhosted.org/packages/bb/e6/9b3afbbcf10cc724312e824af94a2e993d8ace22994d823f5c35324cebf5/pyarrow-18.1.0-cp312-cp312-manylinux_2_28_aarch64.whl", hash = "sha256:d4b3d2a34780645bed6414e22dda55a92e0fcd1b8a637fba86800ad737057e33", size = 38618629 }, + { url = "https://files.pythonhosted.org/packages/3a/2e/3b99f8a3d9e0ccae0e961978a0d0089b25fb46ebbcfb5ebae3cca179a5b3/pyarrow-18.1.0-cp312-cp312-manylinux_2_28_x86_64.whl", hash = "sha256:c52f81aa6f6575058d8e2c782bf79d4f9fdc89887f16825ec3a66607a5dd8e30", size = 40078661 }, + { url = "https://files.pythonhosted.org/packages/41/d7/ed85001edfb96200ff606943cff71d64f91926ab42828676c0fc0db98963/pyarrow-18.1.0-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:acb7564204d3c40babf93a05624fc6a8ec1ab1def295c363afc40b0c9e66c191", size = 39194527 }, + { url = "https://files.pythonhosted.org/packages/59/16/35e28eab126342fa391593415d79477e89582de411bb95232f28b131a769/pyarrow-18.1.0-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:74de649d1d2ccb778f7c3afff6085bd5092aed4c23df9feeb45dd6b16f3811aa", size = 40131443 }, + { url = "https://files.pythonhosted.org/packages/0c/95/e855880614c8da20f4cd74fa85d7268c725cf0013dc754048593a38896a0/pyarrow-18.1.0-cp313-cp313-manylinux_2_28_aarch64.whl", hash = "sha256:f96bd502cb11abb08efea6dab09c003305161cb6c9eafd432e35e76e7fa9b90c", size = 38608750 }, + { url = "https://files.pythonhosted.org/packages/54/9d/f253554b1457d4fdb3831b7bd5f8f00f1795585a606eabf6fec0a58a9c38/pyarrow-18.1.0-cp313-cp313-manylinux_2_28_x86_64.whl", hash = "sha256:36ac22d7782554754a3b50201b607d553a8d71b78cdf03b33c1125be4b52397c", size = 40066690 }, + { url = "https://files.pythonhosted.org/packages/5e/e3/3b16c3190f3d71d3b10f6758d2d5f7779ef008c4fd367cedab3ed178a9f7/pyarrow-18.1.0-cp313-cp313t-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:aebc13a11ed3032d8dd6e7171eb6e86d40d67a5639d96c35142bd568b9299324", size = 39119106 }, + { url = "https://files.pythonhosted.org/packages/1d/d6/5d704b0d25c3c79532f8c0639f253ec2803b897100f64bcb3f53ced236e5/pyarrow-18.1.0-cp313-cp313t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d6cf5c05f3cee251d80e98726b5c7cc9f21bab9e9783673bac58e6dfab57ecc8", size = 40090940 }, + { url = "https://files.pythonhosted.org/packages/37/29/366bc7e588220d74ec00e497ac6710c2833c9176f0372fe0286929b2d64c/pyarrow-18.1.0-cp313-cp313t-manylinux_2_28_aarch64.whl", hash = "sha256:11b676cd410cf162d3f6a70b43fb9e1e40affbc542a1e9ed3681895f2962d3d9", size = 38548177 }, + { url = "https://files.pythonhosted.org/packages/c8/11/fabf6ecabb1fe5b7d96889228ca2a9158c4c3bb732e3b8ee3f7f6d40b703/pyarrow-18.1.0-cp313-cp313t-manylinux_2_28_x86_64.whl", hash = "sha256:b76130d835261b38f14fc41fdfb39ad8d672afb84c447126b84d5472244cfaba", size = 40043567 }, + { url = "https://files.pythonhosted.org/packages/84/c9/62ef9c6281c0e5b4ee1afa9d7bd556e72e06da6706b7906c32c15e69b3d6/pyarrow-18.1.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4f97b31b4c4e21ff58c6f330235ff893cc81e23da081b1a4b1c982075e0ed4e9", size = 39226870 }, + { url = "https://files.pythonhosted.org/packages/b2/99/a6e89e71655a38475e76b060777c8bf69c078b772bec3b7daf7361440f05/pyarrow-18.1.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4a4813cb8ecf1809871fd2d64a8eff740a1bd3691bbe55f01a3cf6c5ec869754", size = 40139114 }, + { url = "https://files.pythonhosted.org/packages/64/a9/06d79923890682e4fe7a16524abee307407008a413115354aaf3226b8410/pyarrow-18.1.0-cp39-cp39-manylinux_2_28_aarch64.whl", hash = "sha256:05a5636ec3eb5cc2a36c6edb534a38ef57b2ab127292a716d00eabb887835f1e", size = 38639231 }, + { url = "https://files.pythonhosted.org/packages/3b/8c/4c3ed19026a00740b81fe1c87f3ff235b2763a0a1ddf5711a9d026b775ce/pyarrow-18.1.0-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:73eeed32e724ea3568bb06161cad5fa7751e45bc2228e33dcb10c614044165c7", size = 40070949 }, +] + +[[package]] +name = "pycparser" +version = "2.22" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/1d/b2/31537cf4b1ca988837256c910a668b553fceb8f069bedc4b1c826024b52c/pycparser-2.22.tar.gz", hash = "sha256:491c8be9c040f5390f5bf44a5b07752bd07f56edf992381b05c701439eec10f6", size = 172736 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/13/a3/a812df4e2dd5696d1f351d58b8fe16a405b234ad2886a0dab9183fb78109/pycparser-2.22-py3-none-any.whl", hash = "sha256:c3702b6d3dd8c7abc1afa565d7e63d53a1d0bd86cdc24edd75470f4de499cfcc", size = 117552 }, +] + [[package]] name = "pydantic" -version = "2.10.3" +version = "2.10.4" source = { registry = "https://pypi.org/simple" } dependencies = [ { name = "annotated-types", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, { name = "pydantic-core", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, { name = "typing-extensions", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, ] -sdist = { url = "https://files.pythonhosted.org/packages/45/0f/27908242621b14e649a84e62b133de45f84c255eecb350ab02979844a788/pydantic-2.10.3.tar.gz", hash = "sha256:cb5ac360ce894ceacd69c403187900a02c4b20b693a9dd1d643e1effab9eadf9", size = 786486 } +sdist = { url = "https://files.pythonhosted.org/packages/70/7e/fb60e6fee04d0ef8f15e4e01ff187a196fa976eb0f0ab524af4599e5754c/pydantic-2.10.4.tar.gz", hash = "sha256:82f12e9723da6de4fe2ba888b5971157b3be7ad914267dea8f05f82b28254f06", size = 762094 } wheels = [ - { url = "https://files.pythonhosted.org/packages/62/51/72c18c55cf2f46ff4f91ebcc8f75aa30f7305f3d726be3f4ebffb4ae972b/pydantic-2.10.3-py3-none-any.whl", hash = "sha256:be04d85bbc7b65651c5f8e6b9976ed9c6f41782a55524cef079a34a0bb82144d", size = 456997 }, + { url = "https://files.pythonhosted.org/packages/f3/26/3e1bbe954fde7ee22a6e7d31582c642aad9e84ffe4b5fb61e63b87cd326f/pydantic-2.10.4-py3-none-any.whl", hash = "sha256:597e135ea68be3a37552fb524bc7d0d66dcf93d395acd93a00682f1efcb8ee3d", size = 431765 }, ] [[package]] name = "pydantic-core" -version = "2.27.1" +version = "2.27.2" source = { registry = "https://pypi.org/simple" } dependencies = [ { name = "typing-extensions", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, ] -sdist = { url = "https://files.pythonhosted.org/packages/a6/9f/7de1f19b6aea45aeb441838782d68352e71bfa98ee6fa048d5041991b33e/pydantic_core-2.27.1.tar.gz", hash = "sha256:62a763352879b84aa31058fc931884055fd75089cccbd9d58bb6afd01141b235", size = 412785 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/01/72/59a70165eabbc93b1111d42df9ca016a4aa109409db04304829377947028/pydantic_core-2.27.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:121ceb0e822f79163dd4699e4c54f5ad38b157084d97b34de8b232bcaad70278", size = 1831953 }, - { url = "https://files.pythonhosted.org/packages/7c/0c/24841136476adafd26f94b45bb718a78cb0500bd7b4f8d667b67c29d7b0d/pydantic_core-2.27.1-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:4603137322c18eaf2e06a4495f426aa8d8388940f3c457e7548145011bb68e05", size = 1856071 }, - { url = "https://files.pythonhosted.org/packages/53/5e/c32957a09cceb2af10d7642df45d1e3dbd8596061f700eac93b801de53c0/pydantic_core-2.27.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:a33cd6ad9017bbeaa9ed78a2e0752c5e250eafb9534f308e7a5f7849b0b1bfb4", size = 2038439 }, - { url = "https://files.pythonhosted.org/packages/e4/8f/979ab3eccd118b638cd6d8f980fea8794f45018255a36044dea40fe579d4/pydantic_core-2.27.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:15cc53a3179ba0fcefe1e3ae50beb2784dede4003ad2dfd24f81bba4b23a454f", size = 2787416 }, - { url = "https://files.pythonhosted.org/packages/02/1d/00f2e4626565b3b6d3690dab4d4fe1a26edd6a20e53749eb21ca892ef2df/pydantic_core-2.27.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:45d9c5eb9273aa50999ad6adc6be5e0ecea7e09dbd0d31bd0c65a55a2592ca08", size = 2134548 }, - { url = "https://files.pythonhosted.org/packages/9d/46/3112621204128b90898adc2e721a3cd6cf5626504178d6f32c33b5a43b79/pydantic_core-2.27.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:8bf7b66ce12a2ac52d16f776b31d16d91033150266eb796967a7e4621707e4f6", size = 1989882 }, - { url = "https://files.pythonhosted.org/packages/49/ec/557dd4ff5287ffffdf16a31d08d723de6762bb1b691879dc4423392309bc/pydantic_core-2.27.1-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:655d7dd86f26cb15ce8a431036f66ce0318648f8853d709b4167786ec2fa4807", size = 1995829 }, - { url = "https://files.pythonhosted.org/packages/6e/b2/610dbeb74d8d43921a7234555e4c091cb050a2bdb8cfea86d07791ce01c5/pydantic_core-2.27.1-cp310-cp310-musllinux_1_1_armv7l.whl", hash = "sha256:5556470f1a2157031e676f776c2bc20acd34c1990ca5f7e56f1ebf938b9ab57c", size = 2091257 }, - { url = "https://files.pythonhosted.org/packages/8c/7f/4bf8e9d26a9118521c80b229291fa9558a07cdd9a968ec2d5c1026f14fbc/pydantic_core-2.27.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:f69ed81ab24d5a3bd93861c8c4436f54afdf8e8cc421562b0c7504cf3be58206", size = 2143894 }, - { url = "https://files.pythonhosted.org/packages/01/de/df51b3bac9820d38371f5a261020f505025df732ce566c2a2e7970b84c8c/pydantic_core-2.27.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7f7059ca8d64fea7f238994c97d91f75965216bcbe5f695bb44f354893f11d52", size = 1829727 }, - { url = "https://files.pythonhosted.org/packages/5f/d9/c01d19da8f9e9fbdb2bf99f8358d145a312590374d0dc9dd8dbe484a9cde/pydantic_core-2.27.1-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:bed0f8a0eeea9fb72937ba118f9db0cb7e90773462af7962d382445f3005e5a4", size = 1854282 }, - { url = "https://files.pythonhosted.org/packages/5f/84/7db66eb12a0dc88c006abd6f3cbbf4232d26adfd827a28638c540d8f871d/pydantic_core-2.27.1-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:a3cb37038123447cf0f3ea4c74751f6a9d7afef0eb71aa07bf5f652b5e6a132c", size = 2037437 }, - { url = "https://files.pythonhosted.org/packages/34/ac/a2537958db8299fbabed81167d58cc1506049dba4163433524e06a7d9f4c/pydantic_core-2.27.1-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:84286494f6c5d05243456e04223d5a9417d7f443c3b76065e75001beb26f88de", size = 2780899 }, - { url = "https://files.pythonhosted.org/packages/4a/c1/3e38cd777ef832c4fdce11d204592e135ddeedb6c6f525478a53d1c7d3e5/pydantic_core-2.27.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:acc07b2cfc5b835444b44a9956846b578d27beeacd4b52e45489e93276241025", size = 2135022 }, - { url = "https://files.pythonhosted.org/packages/7a/69/b9952829f80fd555fe04340539d90e000a146f2a003d3fcd1e7077c06c71/pydantic_core-2.27.1-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:4fefee876e07a6e9aad7a8c8c9f85b0cdbe7df52b8a9552307b09050f7512c7e", size = 1987969 }, - { url = "https://files.pythonhosted.org/packages/05/72/257b5824d7988af43460c4e22b63932ed651fe98804cc2793068de7ec554/pydantic_core-2.27.1-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:258c57abf1188926c774a4c94dd29237e77eda19462e5bb901d88adcab6af919", size = 1994625 }, - { url = "https://files.pythonhosted.org/packages/73/c3/78ed6b7f3278a36589bcdd01243189ade7fc9b26852844938b4d7693895b/pydantic_core-2.27.1-cp311-cp311-musllinux_1_1_armv7l.whl", hash = "sha256:35c14ac45fcfdf7167ca76cc80b2001205a8d5d16d80524e13508371fb8cdd9c", size = 2090089 }, - { url = "https://files.pythonhosted.org/packages/8d/c8/b4139b2f78579960353c4cd987e035108c93a78371bb19ba0dc1ac3b3220/pydantic_core-2.27.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:d1b26e1dff225c31897696cab7d4f0a315d4c0d9e8666dbffdb28216f3b17fdc", size = 2142496 }, - { url = "https://files.pythonhosted.org/packages/2c/9d/e1d6c4561d262b52e41b17a7ef8301e2ba80b61e32e94520271029feb5d8/pydantic_core-2.27.1-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:15aae984e46de8d376df515f00450d1522077254ef6b7ce189b38ecee7c9677c", size = 1828096 }, - { url = "https://files.pythonhosted.org/packages/be/65/80ff46de4266560baa4332ae3181fffc4488ea7d37282da1a62d10ab89a4/pydantic_core-2.27.1-cp312-cp312-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:1ba5e3963344ff25fc8c40da90f44b0afca8cfd89d12964feb79ac1411a260ac", size = 1857708 }, - { url = "https://files.pythonhosted.org/packages/d5/ca/3370074ad758b04d9562b12ecdb088597f4d9d13893a48a583fb47682cdf/pydantic_core-2.27.1-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:992cea5f4f3b29d6b4f7f1726ed8ee46c8331c6b4eed6db5b40134c6fe1768bb", size = 2037751 }, - { url = "https://files.pythonhosted.org/packages/b1/e2/4ab72d93367194317b99d051947c071aef6e3eb95f7553eaa4208ecf9ba4/pydantic_core-2.27.1-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:0325336f348dbee6550d129b1627cb8f5351a9dc91aad141ffb96d4937bd9529", size = 2733863 }, - { url = "https://files.pythonhosted.org/packages/8a/c6/8ae0831bf77f356bb73127ce5a95fe115b10f820ea480abbd72d3cc7ccf3/pydantic_core-2.27.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7597c07fbd11515f654d6ece3d0e4e5093edc30a436c63142d9a4b8e22f19c35", size = 2161161 }, - { url = "https://files.pythonhosted.org/packages/f1/f4/b2fe73241da2429400fc27ddeaa43e35562f96cf5b67499b2de52b528cad/pydantic_core-2.27.1-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:3bbd5d8cc692616d5ef6fbbbd50dbec142c7e6ad9beb66b78a96e9c16729b089", size = 1993294 }, - { url = "https://files.pythonhosted.org/packages/77/29/4bb008823a7f4cc05828198153f9753b3bd4c104d93b8e0b1bfe4e187540/pydantic_core-2.27.1-cp312-cp312-musllinux_1_1_aarch64.whl", hash = "sha256:dc61505e73298a84a2f317255fcc72b710b72980f3a1f670447a21efc88f8381", size = 2001468 }, - { url = "https://files.pythonhosted.org/packages/f2/a9/0eaceeba41b9fad851a4107e0cf999a34ae8f0d0d1f829e2574f3d8897b0/pydantic_core-2.27.1-cp312-cp312-musllinux_1_1_armv7l.whl", hash = "sha256:e1f735dc43da318cad19b4173dd1ffce1d84aafd6c9b782b3abc04a0d5a6f5bb", size = 2091413 }, - { url = "https://files.pythonhosted.org/packages/d8/36/eb8697729725bc610fd73940f0d860d791dc2ad557faaefcbb3edbd2b349/pydantic_core-2.27.1-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:f4e5658dbffe8843a0f12366a4c2d1c316dbe09bb4dfbdc9d2d9cd6031de8aae", size = 2154735 }, - { url = "https://files.pythonhosted.org/packages/41/c2/491b59e222ec7e72236e512108ecad532c7f4391a14e971c963f624f7569/pydantic_core-2.27.1-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:42b0e23f119b2b456d07ca91b307ae167cc3f6c846a7b169fca5326e32fdc6cf", size = 1827854 }, - { url = "https://files.pythonhosted.org/packages/e3/f3/363652651779113189cefdbbb619b7b07b7a67ebb6840325117cc8cc3460/pydantic_core-2.27.1-cp313-cp313-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:764be71193f87d460a03f1f7385a82e226639732214b402f9aa61f0d025f0737", size = 1857389 }, - { url = "https://files.pythonhosted.org/packages/5f/97/be804aed6b479af5a945daec7538d8bf358d668bdadde4c7888a2506bdfb/pydantic_core-2.27.1-cp313-cp313-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:1c00666a3bd2f84920a4e94434f5974d7bbc57e461318d6bb34ce9cdbbc1f6b2", size = 2037934 }, - { url = "https://files.pythonhosted.org/packages/42/01/295f0bd4abf58902917e342ddfe5f76cf66ffabfc57c2e23c7681a1a1197/pydantic_core-2.27.1-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:3ccaa88b24eebc0f849ce0a4d09e8a408ec5a94afff395eb69baf868f5183107", size = 2735176 }, - { url = "https://files.pythonhosted.org/packages/9d/a0/cd8e9c940ead89cc37812a1a9f310fef59ba2f0b22b4e417d84ab09fa970/pydantic_core-2.27.1-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c65af9088ac534313e1963443d0ec360bb2b9cba6c2909478d22c2e363d98a51", size = 2160720 }, - { url = "https://files.pythonhosted.org/packages/73/ae/9d0980e286627e0aeca4c352a60bd760331622c12d576e5ea4441ac7e15e/pydantic_core-2.27.1-cp313-cp313-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:206b5cf6f0c513baffaeae7bd817717140770c74528f3e4c3e1cec7871ddd61a", size = 1992972 }, - { url = "https://files.pythonhosted.org/packages/bf/ba/ae4480bc0292d54b85cfb954e9d6bd226982949f8316338677d56541b85f/pydantic_core-2.27.1-cp313-cp313-musllinux_1_1_aarch64.whl", hash = "sha256:062f60e512fc7fff8b8a9d680ff0ddaaef0193dba9fa83e679c0c5f5fbd018bc", size = 2001477 }, - { url = "https://files.pythonhosted.org/packages/55/b7/e26adf48c2f943092ce54ae14c3c08d0d221ad34ce80b18a50de8ed2cba8/pydantic_core-2.27.1-cp313-cp313-musllinux_1_1_armv7l.whl", hash = "sha256:a0697803ed7d4af5e4c1adf1670af078f8fcab7a86350e969f454daf598c4960", size = 2091186 }, - { url = "https://files.pythonhosted.org/packages/ba/cc/8491fff5b608b3862eb36e7d29d36a1af1c945463ca4c5040bf46cc73f40/pydantic_core-2.27.1-cp313-cp313-musllinux_1_1_x86_64.whl", hash = "sha256:58ca98a950171f3151c603aeea9303ef6c235f692fe555e883591103da709b23", size = 2154429 }, - { url = "https://files.pythonhosted.org/packages/36/b8/6f1b7c5f068c00dfe179b8762bc1d32c75c0e9f62c9372174b1b64a74aa8/pydantic_core-2.27.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0b3dfe500de26c52abe0477dde16192ac39c98f05bf2d80e76102d394bd13854", size = 1832148 }, - { url = "https://files.pythonhosted.org/packages/d9/83/83ff64d599847f080a93df119e856e3bd93063cced04b9a27eb66d863831/pydantic_core-2.27.1-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:aee66be87825cdf72ac64cb03ad4c15ffef4143dbf5c113f64a5ff4f81477bf9", size = 1856371 }, - { url = "https://files.pythonhosted.org/packages/72/e9/974e6c73f59627c446833ecc306cadd199edab40abcfa093372a5a5c0156/pydantic_core-2.27.1-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3b748c44bb9f53031c8cbc99a8a061bc181c1000c60a30f55393b6e9c45cc5bd", size = 2038686 }, - { url = "https://files.pythonhosted.org/packages/5e/bb/5e912d02dcf29aebb2da35e5a1a26088c39ffc0b1ea81242ee9db6f1f730/pydantic_core-2.27.1-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:5ca038c7f6a0afd0b2448941b6ef9d5e1949e999f9e5517692eb6da58e9d44be", size = 2785725 }, - { url = "https://files.pythonhosted.org/packages/85/d7/936846087424c882d89c853711687230cd60179a67c79c34c99b64f92625/pydantic_core-2.27.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6e0bd57539da59a3e4671b90a502da9a28c72322a4f17866ba3ac63a82c4498e", size = 2135177 }, - { url = "https://files.pythonhosted.org/packages/82/72/5a386e5ce8d3e933c3f283e61357474181c39383f38afffc15a6152fa1c5/pydantic_core-2.27.1-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:ac6c2c45c847bbf8f91930d88716a0fb924b51e0c6dad329b793d670ec5db792", size = 1989877 }, - { url = "https://files.pythonhosted.org/packages/ce/5c/b1c417a5fd67ce132d78d16a6ba7629dc7f188dbd4f7c30ef58111ee5147/pydantic_core-2.27.1-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:b94d4ba43739bbe8b0ce4262bcc3b7b9f31459ad120fb595627eaeb7f9b9ca01", size = 1996006 }, - { url = "https://files.pythonhosted.org/packages/dd/04/4e18f2c42b29929882f30e4c09a3a039555158995a4ac730a73585198a66/pydantic_core-2.27.1-cp39-cp39-musllinux_1_1_armv7l.whl", hash = "sha256:00e6424f4b26fe82d44577b4c842d7df97c20be6439e8e685d0d715feceb9fb9", size = 2091441 }, - { url = "https://files.pythonhosted.org/packages/06/84/5a332345b7efb5ab361f916eaf7316ef010e72417e8c7dd3d34462ee9840/pydantic_core-2.27.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:38de0a70160dd97540335b7ad3a74571b24f1dc3ed33f815f0880682e6880131", size = 2144471 }, - { url = "https://files.pythonhosted.org/packages/8f/fc/5485cf0b0bb38da31d1d292160a4d123b5977841ddc1122c671a30b76cfd/pydantic_core-2.27.1-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0e4216e64d203e39c62df627aa882f02a2438d18a5f21d7f721621f7a5d3611d", size = 1826929 }, - { url = "https://files.pythonhosted.org/packages/a1/ff/fb1284a210e13a5f34c639efc54d51da136074ffbe25ec0c279cf9fbb1c4/pydantic_core-2.27.1-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:02a3d637bd387c41d46b002f0e49c52642281edacd2740e5a42f7017feea3f2c", size = 1980591 }, - { url = "https://files.pythonhosted.org/packages/f1/14/77c1887a182d05af74f6aeac7b740da3a74155d3093ccc7ee10b900cc6b5/pydantic_core-2.27.1-pp310-pypy310_pp73-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:161c27ccce13b6b0c8689418da3885d3220ed2eae2ea5e9b2f7f3d48f1d52c27", size = 1981326 }, - { url = "https://files.pythonhosted.org/packages/06/aa/6f1b2747f811a9c66b5ef39d7f02fbb200479784c75e98290d70004b1253/pydantic_core-2.27.1-pp310-pypy310_pp73-musllinux_1_1_aarch64.whl", hash = "sha256:19910754e4cc9c63bc1c7f6d73aa1cfee82f42007e407c0f413695c2f7ed777f", size = 1989205 }, - { url = "https://files.pythonhosted.org/packages/7a/d2/8ce2b074d6835f3c88d85f6d8a399790043e9fdb3d0e43455e72d19df8cc/pydantic_core-2.27.1-pp310-pypy310_pp73-musllinux_1_1_armv7l.whl", hash = "sha256:e173486019cc283dc9778315fa29a363579372fe67045e971e89b6365cc035ed", size = 2079616 }, - { url = "https://files.pythonhosted.org/packages/65/71/af01033d4e58484c3db1e5d13e751ba5e3d6b87cc3368533df4c50932c8b/pydantic_core-2.27.1-pp310-pypy310_pp73-musllinux_1_1_x86_64.whl", hash = "sha256:af52d26579b308921b73b956153066481f064875140ccd1dfd4e77db89dbb12f", size = 2133265 }, - { url = "https://files.pythonhosted.org/packages/7f/ee/0377e9f4ca5a47e8885f670a65c0a647ddf9ce98d50bf7547cf8e1ee5771/pydantic_core-2.27.1-pp39-pypy39_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9c10c309e18e443ddb108f0ef64e8729363adbfd92d6d57beec680f6261556f3", size = 1827025 }, - { url = "https://files.pythonhosted.org/packages/fe/0b/a24d9ef762d05bebdfafd6d5d176b990728fa9ec8ea7b6040d6fb5f3caaa/pydantic_core-2.27.1-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:98476c98b02c8e9b2eec76ac4156fd006628b1b2d0ef27e548ffa978393fd154", size = 1980927 }, - { url = "https://files.pythonhosted.org/packages/00/bd/deadc1722eb7dfdf787a3bbcd32eabbdcc36931fd48671a850e1b9f2cd77/pydantic_core-2.27.1-pp39-pypy39_pp73-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:c3027001c28434e7ca5a6e1e527487051136aa81803ac812be51802150d880dd", size = 1980918 }, - { url = "https://files.pythonhosted.org/packages/f0/05/5d09d0b0e92053d538927308ea1d35cb25ab543d9c3e2eb2d7653bc73690/pydantic_core-2.27.1-pp39-pypy39_pp73-musllinux_1_1_aarch64.whl", hash = "sha256:7699b1df36a48169cdebda7ab5a2bac265204003f153b4bd17276153d997670a", size = 1989990 }, - { url = "https://files.pythonhosted.org/packages/5b/7e/f7191346d1c3ac66049f618ee331359f8552a8b68a2daf916003c30b6dc8/pydantic_core-2.27.1-pp39-pypy39_pp73-musllinux_1_1_armv7l.whl", hash = "sha256:1c39b07d90be6b48968ddc8c19e7585052088fd7ec8d568bb31ff64c70ae3c97", size = 2079871 }, - { url = "https://files.pythonhosted.org/packages/f3/65/2caf4f7ad65413a137d43cb9578c54d1abd3224be786ad840263c1bf9e0f/pydantic_core-2.27.1-pp39-pypy39_pp73-musllinux_1_1_x86_64.whl", hash = "sha256:46ccfe3032b3915586e469d4972973f893c0a2bb65669194a5bdea9bacc088c2", size = 2133569 }, +sdist = { url = "https://files.pythonhosted.org/packages/fc/01/f3e5ac5e7c25833db5eb555f7b7ab24cd6f8c322d3a3ad2d67a952dc0abc/pydantic_core-2.27.2.tar.gz", hash = "sha256:eb026e5a4c1fee05726072337ff51d1efb6f59090b7da90d30ea58625b1ffb39", size = 413443 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/c3/43/fafabd3d94d159d4f1ed62e383e264f146a17dd4d48453319fd782e7979e/pydantic_core-2.27.2-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7969e133a6f183be60e9f6f56bfae753585680f3b7307a8e555a948d443cc05a", size = 1829169 }, + { url = "https://files.pythonhosted.org/packages/a2/d1/f2dfe1a2a637ce6800b799aa086d079998959f6f1215eb4497966efd2274/pydantic_core-2.27.2-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:3de9961f2a346257caf0aa508a4da705467f53778e9ef6fe744c038119737ef5", size = 1867227 }, + { url = "https://files.pythonhosted.org/packages/7d/39/e06fcbcc1c785daa3160ccf6c1c38fea31f5754b756e34b65f74e99780b5/pydantic_core-2.27.2-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e2bb4d3e5873c37bb3dd58714d4cd0b0e6238cebc4177ac8fe878f8b3aa8e74c", size = 2037695 }, + { url = "https://files.pythonhosted.org/packages/7a/67/61291ee98e07f0650eb756d44998214231f50751ba7e13f4f325d95249ab/pydantic_core-2.27.2-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:280d219beebb0752699480fe8f1dc61ab6615c2046d76b7ab7ee38858de0a4e7", size = 2741662 }, + { url = "https://files.pythonhosted.org/packages/32/90/3b15e31b88ca39e9e626630b4c4a1f5a0dfd09076366f4219429e6786076/pydantic_core-2.27.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:47956ae78b6422cbd46f772f1746799cbb862de838fd8d1fbd34a82e05b0983a", size = 1993370 }, + { url = "https://files.pythonhosted.org/packages/ff/83/c06d333ee3a67e2e13e07794995c1535565132940715931c1c43bfc85b11/pydantic_core-2.27.2-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:14d4a5c49d2f009d62a2a7140d3064f686d17a5d1a268bc641954ba181880236", size = 1996813 }, + { url = "https://files.pythonhosted.org/packages/7c/f7/89be1c8deb6e22618a74f0ca0d933fdcb8baa254753b26b25ad3acff8f74/pydantic_core-2.27.2-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:337b443af21d488716f8d0b6164de833e788aa6bd7e3a39c005febc1284f4962", size = 2005287 }, + { url = "https://files.pythonhosted.org/packages/b7/7d/8eb3e23206c00ef7feee17b83a4ffa0a623eb1a9d382e56e4aa46fd15ff2/pydantic_core-2.27.2-cp310-cp310-musllinux_1_1_armv7l.whl", hash = "sha256:03d0f86ea3184a12f41a2d23f7ccb79cdb5a18e06993f8a45baa8dfec746f0e9", size = 2128414 }, + { url = "https://files.pythonhosted.org/packages/4e/99/fe80f3ff8dd71a3ea15763878d464476e6cb0a2db95ff1c5c554133b6b83/pydantic_core-2.27.2-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:7041c36f5680c6e0f08d922aed302e98b3745d97fe1589db0a3eebf6624523af", size = 2155301 }, + { url = "https://files.pythonhosted.org/packages/a6/3c/724039e0d848fd69dbf5806894e26479577316c6f0f112bacaf67aa889ac/pydantic_core-2.27.2-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8c19d1ea0673cd13cc2f872f6c9ab42acc4e4f492a7ca9d3795ce2b112dd7e15", size = 1826167 }, + { url = "https://files.pythonhosted.org/packages/2b/5b/1b29e8c1fb5f3199a9a57c1452004ff39f494bbe9bdbe9a81e18172e40d3/pydantic_core-2.27.2-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:5e68c4446fe0810e959cdff46ab0a41ce2f2c86d227d96dc3847af0ba7def306", size = 1865071 }, + { url = "https://files.pythonhosted.org/packages/89/6c/3985203863d76bb7d7266e36970d7e3b6385148c18a68cc8915fd8c84d57/pydantic_core-2.27.2-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d9640b0059ff4f14d1f37321b94061c6db164fbe49b334b31643e0528d100d99", size = 2036244 }, + { url = "https://files.pythonhosted.org/packages/0e/41/f15316858a246b5d723f7d7f599f79e37493b2e84bfc789e58d88c209f8a/pydantic_core-2.27.2-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:40d02e7d45c9f8af700f3452f329ead92da4c5f4317ca9b896de7ce7199ea459", size = 2737470 }, + { url = "https://files.pythonhosted.org/packages/a8/7c/b860618c25678bbd6d1d99dbdfdf0510ccb50790099b963ff78a124b754f/pydantic_core-2.27.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1c1fd185014191700554795c99b347d64f2bb637966c4cfc16998a0ca700d048", size = 1992291 }, + { url = "https://files.pythonhosted.org/packages/bf/73/42c3742a391eccbeab39f15213ecda3104ae8682ba3c0c28069fbcb8c10d/pydantic_core-2.27.2-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:d81d2068e1c1228a565af076598f9e7451712700b673de8f502f0334f281387d", size = 1994613 }, + { url = "https://files.pythonhosted.org/packages/94/7a/941e89096d1175d56f59340f3a8ebaf20762fef222c298ea96d36a6328c5/pydantic_core-2.27.2-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:1a4207639fb02ec2dbb76227d7c751a20b1a6b4bc52850568e52260cae64ca3b", size = 2002355 }, + { url = "https://files.pythonhosted.org/packages/6e/95/2359937a73d49e336a5a19848713555605d4d8d6940c3ec6c6c0ca4dcf25/pydantic_core-2.27.2-cp311-cp311-musllinux_1_1_armv7l.whl", hash = "sha256:3de3ce3c9ddc8bbd88f6e0e304dea0e66d843ec9de1b0042b0911c1663ffd474", size = 2126661 }, + { url = "https://files.pythonhosted.org/packages/2b/4c/ca02b7bdb6012a1adef21a50625b14f43ed4d11f1fc237f9d7490aa5078c/pydantic_core-2.27.2-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:30c5f68ded0c36466acede341551106821043e9afaad516adfb6e8fa80a4e6a6", size = 2153261 }, + { url = "https://files.pythonhosted.org/packages/9e/91/840ec1375e686dbae1bd80a9e46c26a1e0083e1186abc610efa3d9a36180/pydantic_core-2.27.2-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:172fce187655fece0c90d90a678424b013f8fbb0ca8b036ac266749c09438cb7", size = 1822900 }, + { url = "https://files.pythonhosted.org/packages/f6/31/4240bc96025035500c18adc149aa6ffdf1a0062a4b525c932065ceb4d868/pydantic_core-2.27.2-cp312-cp312-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:519f29f5213271eeeeb3093f662ba2fd512b91c5f188f3bb7b27bc5973816934", size = 1869177 }, + { url = "https://files.pythonhosted.org/packages/fa/20/02fbaadb7808be578317015c462655c317a77a7c8f0ef274bc016a784c54/pydantic_core-2.27.2-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:05e3a55d124407fffba0dd6b0c0cd056d10e983ceb4e5dbd10dda135c31071d6", size = 2038046 }, + { url = "https://files.pythonhosted.org/packages/06/86/7f306b904e6c9eccf0668248b3f272090e49c275bc488a7b88b0823444a4/pydantic_core-2.27.2-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9c3ed807c7b91de05e63930188f19e921d1fe90de6b4f5cd43ee7fcc3525cb8c", size = 2685386 }, + { url = "https://files.pythonhosted.org/packages/8d/f0/49129b27c43396581a635d8710dae54a791b17dfc50c70164866bbf865e3/pydantic_core-2.27.2-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6fb4aadc0b9a0c063206846d603b92030eb6f03069151a625667f982887153e2", size = 1997060 }, + { url = "https://files.pythonhosted.org/packages/0d/0f/943b4af7cd416c477fd40b187036c4f89b416a33d3cc0ab7b82708a667aa/pydantic_core-2.27.2-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:28ccb213807e037460326424ceb8b5245acb88f32f3d2777427476e1b32c48c4", size = 2004870 }, + { url = "https://files.pythonhosted.org/packages/35/40/aea70b5b1a63911c53a4c8117c0a828d6790483f858041f47bab0b779f44/pydantic_core-2.27.2-cp312-cp312-musllinux_1_1_aarch64.whl", hash = "sha256:de3cd1899e2c279b140adde9357c4495ed9d47131b4a4eaff9052f23398076b3", size = 1999822 }, + { url = "https://files.pythonhosted.org/packages/f2/b3/807b94fd337d58effc5498fd1a7a4d9d59af4133e83e32ae39a96fddec9d/pydantic_core-2.27.2-cp312-cp312-musllinux_1_1_armv7l.whl", hash = "sha256:220f892729375e2d736b97d0e51466252ad84c51857d4d15f5e9692f9ef12be4", size = 2130364 }, + { url = "https://files.pythonhosted.org/packages/fc/df/791c827cd4ee6efd59248dca9369fb35e80a9484462c33c6649a8d02b565/pydantic_core-2.27.2-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:a0fcd29cd6b4e74fe8ddd2c90330fd8edf2e30cb52acda47f06dd615ae72da57", size = 2158303 }, + { url = "https://files.pythonhosted.org/packages/ba/15/52cfe49c8c986e081b863b102d6b859d9defc63446b642ccbbb3742bf371/pydantic_core-2.27.2-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:71b24c7d61131bb83df10cc7e687433609963a944ccf45190cfc21e0887b08c9", size = 1823027 }, + { url = "https://files.pythonhosted.org/packages/b1/1c/b6f402cfc18ec0024120602bdbcebc7bdd5b856528c013bd4d13865ca473/pydantic_core-2.27.2-cp313-cp313-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:fa8e459d4954f608fa26116118bb67f56b93b209c39b008277ace29937453dc9", size = 1868888 }, + { url = "https://files.pythonhosted.org/packages/bd/7b/8cb75b66ac37bc2975a3b7de99f3c6f355fcc4d89820b61dffa8f1e81677/pydantic_core-2.27.2-cp313-cp313-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ce8918cbebc8da707ba805b7fd0b382816858728ae7fe19a942080c24e5b7cd1", size = 2037738 }, + { url = "https://files.pythonhosted.org/packages/c8/f1/786d8fe78970a06f61df22cba58e365ce304bf9b9f46cc71c8c424e0c334/pydantic_core-2.27.2-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:eda3f5c2a021bbc5d976107bb302e0131351c2ba54343f8a496dc8783d3d3a6a", size = 2685138 }, + { url = "https://files.pythonhosted.org/packages/a6/74/d12b2cd841d8724dc8ffb13fc5cef86566a53ed358103150209ecd5d1999/pydantic_core-2.27.2-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bd8086fa684c4775c27f03f062cbb9eaa6e17f064307e86b21b9e0abc9c0f02e", size = 1997025 }, + { url = "https://files.pythonhosted.org/packages/a0/6e/940bcd631bc4d9a06c9539b51f070b66e8f370ed0933f392db6ff350d873/pydantic_core-2.27.2-cp313-cp313-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:8d9b3388db186ba0c099a6d20f0604a44eabdeef1777ddd94786cdae158729e4", size = 2004633 }, + { url = "https://files.pythonhosted.org/packages/50/cc/a46b34f1708d82498c227d5d80ce615b2dd502ddcfd8376fc14a36655af1/pydantic_core-2.27.2-cp313-cp313-musllinux_1_1_aarch64.whl", hash = "sha256:7a66efda2387de898c8f38c0cf7f14fca0b51a8ef0b24bfea5849f1b3c95af27", size = 1999404 }, + { url = "https://files.pythonhosted.org/packages/ca/2d/c365cfa930ed23bc58c41463bae347d1005537dc8db79e998af8ba28d35e/pydantic_core-2.27.2-cp313-cp313-musllinux_1_1_armv7l.whl", hash = "sha256:18a101c168e4e092ab40dbc2503bdc0f62010e95d292b27827871dc85450d7ee", size = 2130130 }, + { url = "https://files.pythonhosted.org/packages/f4/d7/eb64d015c350b7cdb371145b54d96c919d4db516817f31cd1c650cae3b21/pydantic_core-2.27.2-cp313-cp313-musllinux_1_1_x86_64.whl", hash = "sha256:ba5dd002f88b78a4215ed2f8ddbdf85e8513382820ba15ad5ad8955ce0ca19a1", size = 2157946 }, + { url = "https://files.pythonhosted.org/packages/8a/9e/e44b8cb0edf04a2f0a1f6425a65ee089c1d6f9c4c2dcab0209127b6fdfc2/pydantic_core-2.27.2-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c61709a844acc6bf0b7dce7daae75195a10aac96a596ea1b776996414791ede4", size = 1829112 }, + { url = "https://files.pythonhosted.org/packages/1c/90/1160d7ac700102effe11616e8119e268770f2a2aa5afb935f3ee6832987d/pydantic_core-2.27.2-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:42c5f762659e47fdb7b16956c71598292f60a03aa92f8b6351504359dbdba6cf", size = 1866780 }, + { url = "https://files.pythonhosted.org/packages/ee/33/13983426df09a36d22c15980008f8d9c77674fc319351813b5a2739b70f3/pydantic_core-2.27.2-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4c9775e339e42e79ec99c441d9730fccf07414af63eac2f0e48e08fd38a64d76", size = 2037943 }, + { url = "https://files.pythonhosted.org/packages/01/d7/ced164e376f6747e9158c89988c293cd524ab8d215ae4e185e9929655d5c/pydantic_core-2.27.2-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:57762139821c31847cfb2df63c12f725788bd9f04bc2fb392790959b8f70f118", size = 2740492 }, + { url = "https://files.pythonhosted.org/packages/8b/1f/3dc6e769d5b7461040778816aab2b00422427bcaa4b56cc89e9c653b2605/pydantic_core-2.27.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0d1e85068e818c73e048fe28cfc769040bb1f475524f4745a5dc621f75ac7630", size = 1995714 }, + { url = "https://files.pythonhosted.org/packages/07/d7/a0bd09bc39283530b3f7c27033a814ef254ba3bd0b5cfd040b7abf1fe5da/pydantic_core-2.27.2-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:097830ed52fd9e427942ff3b9bc17fab52913b2f50f2880dc4a5611446606a54", size = 1997163 }, + { url = "https://files.pythonhosted.org/packages/2d/bb/2db4ad1762e1c5699d9b857eeb41959191980de6feb054e70f93085e1bcd/pydantic_core-2.27.2-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:044a50963a614ecfae59bb1eaf7ea7efc4bc62f49ed594e18fa1e5d953c40e9f", size = 2005217 }, + { url = "https://files.pythonhosted.org/packages/53/5f/23a5a3e7b8403f8dd8fc8a6f8b49f6b55c7d715b77dcf1f8ae919eeb5628/pydantic_core-2.27.2-cp39-cp39-musllinux_1_1_armv7l.whl", hash = "sha256:4e0b4220ba5b40d727c7f879eac379b822eee5d8fff418e9d3381ee45b3b0362", size = 2127899 }, + { url = "https://files.pythonhosted.org/packages/c2/ae/aa38bb8dd3d89c2f1d8362dd890ee8f3b967330821d03bbe08fa01ce3766/pydantic_core-2.27.2-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:5e4f4bb20d75e9325cc9696c6802657b58bc1dbbe3022f32cc2b2b632c3fbb96", size = 2155726 }, + { url = "https://files.pythonhosted.org/packages/f6/f0/a3ae8fbee269e4934f14e2e0e00928f9346c5943174f2811193113e58252/pydantic_core-2.27.2-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:688d3fd9fcb71f41c4c015c023d12a79d1c4c0732ec9eb35d96e3388a120dcf3", size = 1822467 }, + { url = "https://files.pythonhosted.org/packages/d7/7a/7bbf241a04e9f9ea24cd5874354a83526d639b02674648af3f350554276c/pydantic_core-2.27.2-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3d591580c34f4d731592f0e9fe40f9cc1b430d297eecc70b962e93c5c668f15f", size = 1979797 }, + { url = "https://files.pythonhosted.org/packages/4f/5f/4784c6107731f89e0005a92ecb8a2efeafdb55eb992b8e9d0a2be5199335/pydantic_core-2.27.2-pp310-pypy310_pp73-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:82f986faf4e644ffc189a7f1aafc86e46ef70372bb153e7001e8afccc6e54133", size = 1987839 }, + { url = "https://files.pythonhosted.org/packages/6d/a7/61246562b651dff00de86a5f01b6e4befb518df314c54dec187a78d81c84/pydantic_core-2.27.2-pp310-pypy310_pp73-musllinux_1_1_aarch64.whl", hash = "sha256:bec317a27290e2537f922639cafd54990551725fc844249e64c523301d0822fc", size = 1998861 }, + { url = "https://files.pythonhosted.org/packages/86/aa/837821ecf0c022bbb74ca132e117c358321e72e7f9702d1b6a03758545e2/pydantic_core-2.27.2-pp310-pypy310_pp73-musllinux_1_1_armv7l.whl", hash = "sha256:0296abcb83a797db256b773f45773da397da75a08f5fcaef41f2044adec05f50", size = 2116582 }, + { url = "https://files.pythonhosted.org/packages/81/b0/5e74656e95623cbaa0a6278d16cf15e10a51f6002e3ec126541e95c29ea3/pydantic_core-2.27.2-pp310-pypy310_pp73-musllinux_1_1_x86_64.whl", hash = "sha256:0d75070718e369e452075a6017fbf187f788e17ed67a3abd47fa934d001863d9", size = 2151985 }, + { url = "https://files.pythonhosted.org/packages/41/f7/f847b15fb14978ca2b30262548f5fc4872b2724e90f116393eb69008299d/pydantic_core-2.27.2-pp39-pypy39_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c817e2b40aba42bac6f457498dacabc568c3b7a986fc9ba7c8d9d260b71485fb", size = 1822307 }, + { url = "https://files.pythonhosted.org/packages/9c/63/ed80ec8255b587b2f108e514dc03eed1546cd00f0af281e699797f373f38/pydantic_core-2.27.2-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:251136cdad0cb722e93732cb45ca5299fb56e1344a833640bf93b2803f8d1bfd", size = 1979971 }, + { url = "https://files.pythonhosted.org/packages/a9/6d/6d18308a45454a0de0e975d70171cadaf454bc7a0bf86b9c7688e313f0bb/pydantic_core-2.27.2-pp39-pypy39_pp73-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:d2088237af596f0a524d3afc39ab3b036e8adb054ee57cbb1dcf8e09da5b29cc", size = 1987616 }, + { url = "https://files.pythonhosted.org/packages/82/8a/05f8780f2c1081b800a7ca54c1971e291c2d07d1a50fb23c7e4aef4ed403/pydantic_core-2.27.2-pp39-pypy39_pp73-musllinux_1_1_aarch64.whl", hash = "sha256:d4041c0b966a84b4ae7a09832eb691a35aec90910cd2dbe7a208de59be77965b", size = 1998943 }, + { url = "https://files.pythonhosted.org/packages/5e/3e/fe5b6613d9e4c0038434396b46c5303f5ade871166900b357ada4766c5b7/pydantic_core-2.27.2-pp39-pypy39_pp73-musllinux_1_1_armv7l.whl", hash = "sha256:8083d4e875ebe0b864ffef72a4304827015cff328a1be6e22cc850753bfb122b", size = 2116654 }, + { url = "https://files.pythonhosted.org/packages/db/ad/28869f58938fad8cc84739c4e592989730bfb69b7c90a8fff138dff18e1e/pydantic_core-2.27.2-pp39-pypy39_pp73-musllinux_1_1_x86_64.whl", hash = "sha256:f141ee28a0ad2123b6611b6ceff018039df17f32ada8b534e6aa039545a3efb2", size = 2152292 }, ] [[package]] @@ -949,6 +1792,27 @@ wheels = [ { url = "https://files.pythonhosted.org/packages/f7/3f/01c8b82017c199075f8f788d0d906b9ffbbc5a47dc9918a945e13d5a2bda/pygments-2.18.0-py3-none-any.whl", hash = "sha256:b8e6aca0523f3ab76fee51799c488e38782ac06eafcf95e7ba832985c8e7b13a", size = 1205513 }, ] +[[package]] +name = "pynvml" +version = "12.0.0" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "nvidia-ml-py", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/26/6f/6b5880ed0239e85b9a39aed103b65b2ef81425beef9f45e5c035bf008330/pynvml-12.0.0.tar.gz", hash = "sha256:299ce2451a6a17e6822d6faee750103e25b415f06f59abb8db65d30f794166f5", size = 33636 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/ed/df/f7cf07a65a96dd11d71f346f9c2863accdd4784da83af7181b067d556cbc/pynvml-12.0.0-py3-none-any.whl", hash = "sha256:fdff84b62a27dbe98e08e1a647eb77342bef1aebe0878bcd15e99a83fcbecb9e", size = 26560 }, +] + +[[package]] +name = "pyproject-hooks" +version = "1.2.0" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/e7/82/28175b2414effca1cdac8dc99f76d660e7a4fb0ceefa4b4ab8f5f6742925/pyproject_hooks-1.2.0.tar.gz", hash = "sha256:1e859bd5c40fae9448642dd871adf459e5e2084186e8d2c2a79a824c970da1f8", size = 19228 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/bd/24/12818598c362d7f300f18e74db45963dbcb85150324092410c8b49405e42/pyproject_hooks-1.2.0-py3-none-any.whl", hash = "sha256:9e5c6bfa8dcc30091c74b0cf803c81fdd29d94f01992a7707bc97babb1141913", size = 10216 }, +] + [[package]] name = "pytest" version = "8.3.4" @@ -978,26 +1842,43 @@ wheels = [ { url = "https://files.pythonhosted.org/packages/6d/82/1d96bf03ee4c0fdc3c0cbe61470070e659ca78dc0086fb88b66c185e2449/pytest_xdist-3.6.1-py3-none-any.whl", hash = "sha256:9ed4adfb68a016610848639bb7e02c9352d5d9f03d04809919e2dafc3be4cca7", size = 46108 }, ] +[[package]] +name = "python-dateutil" +version = "2.9.0.post0" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "six", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/66/c0/0c8b6ad9f17a802ee498c46e004a0eb49bc148f2fd230864601a86dcf6db/python-dateutil-2.9.0.post0.tar.gz", hash = "sha256:37dd54208da7e1cd875388217d5e00ebd4179249f90fb72437e91a35459a0ad3", size = 342432 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/ec/57/56b9bcc3c9c6a792fcbaf139543cee77261f3651ca9da0c93f5c1221264b/python_dateutil-2.9.0.post0-py2.py3-none-any.whl", hash = "sha256:a8b2bc7bffae282281c8140a97d3aa9c14da0b136dfe83f850eea9a5f7470427", size = 229892 }, +] + [[package]] name = "pytorch-triton" -version = "3.2.0+git35c6c7c6" +version = "3.2.0+git0d4682f0" source = { registry = "https://download.pytorch.org/whl/nightly/cu126" } wheels = [ - { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit35c6c7c6-cp310-cp310-linux_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit35c6c7c6-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit35c6c7c6-cp310-cp310-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit35c6c7c6-cp311-cp311-linux_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit35c6c7c6-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit35c6c7c6-cp311-cp311-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit35c6c7c6-cp312-cp312-linux_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit35c6c7c6-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit35c6c7c6-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit35c6c7c6-cp313-cp313-linux_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit35c6c7c6-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit35c6c7c6-cp313-cp313-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit35c6c7c6-cp39-cp39-linux_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit35c6c7c6-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit35c6c7c6-cp39-cp39-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit0d4682f0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit0d4682f0-cp310-cp310-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit0d4682f0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit0d4682f0-cp311-cp311-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit0d4682f0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit0d4682f0-cp312-cp312-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit0d4682f0-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit0d4682f0-cp313-cp313-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit0d4682f0-cp313-cp313t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit0d4682f0-cp313-cp313t-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit0d4682f0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/pytorch_triton-3.2.0%2Bgit0d4682f0-cp39-cp39-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl" }, +] + +[[package]] +name = "pytz" +version = "2024.2" +source = { registry = "https://download.pytorch.org/whl/nightly/cu126" } +wheels = [ + { url = "https://download.pytorch.org/whl/nightly/pytz-2024.2-py2.py3-none-any.whl" }, ] [[package]] @@ -1022,6 +1903,65 @@ wheels = [ { url = "https://download.pytorch.org/whl/nightly/PyYAML-6.0.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, ] +[[package]] +name = "pyzmq" +version = "26.2.0" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "cffi", marker = "(implementation_name == 'pypy' and sys_platform == 'linux') or (implementation_name == 'pypy' and sys_platform == 'windows')" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/fd/05/bed626b9f7bb2322cdbbf7b4bd8f54b1b617b0d2ab2d3547d6e39428a48e/pyzmq-26.2.0.tar.gz", hash = "sha256:070672c258581c8e4f640b5159297580a9974b026043bd4ab0470be9ed324f1f", size = 271975 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/b6/09/b51b6683fde5ca04593a57bbe81788b6b43114d8f8ee4e80afc991e14760/pyzmq-26.2.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:89289a5ee32ef6c439086184529ae060c741334b8970a6855ec0b6ad3ff28764", size = 673199 }, + { url = "https://files.pythonhosted.org/packages/c9/78/486f3e2e824f3a645238332bf5a4c4b4477c3063033a27c1e4052358dee2/pyzmq-26.2.0-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5506f06d7dc6ecf1efacb4a013b1f05071bb24b76350832c96449f4a2d95091c", size = 911762 }, + { url = "https://files.pythonhosted.org/packages/5e/3b/2eb1667c9b866f53e76ee8b0c301b0469745a23bd5a87b7ee3d5dd9eb6e5/pyzmq-26.2.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8ea039387c10202ce304af74def5021e9adc6297067f3441d348d2b633e8166a", size = 868773 }, + { url = "https://files.pythonhosted.org/packages/16/29/ca99b4598a9dc7e468b5417eda91f372b595be1e3eec9b7cbe8e5d3584e8/pyzmq-26.2.0-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:a2224fa4a4c2ee872886ed00a571f5e967c85e078e8e8c2530a2fb01b3309b88", size = 868834 }, + { url = "https://files.pythonhosted.org/packages/ad/e5/9efaeb1d2f4f8c50da04144f639b042bc52869d3a206d6bf672ab3522163/pyzmq-26.2.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:28ad5233e9c3b52d76196c696e362508959741e1a005fb8fa03b51aea156088f", size = 1202861 }, + { url = "https://files.pythonhosted.org/packages/c3/62/c721b5608a8ac0a69bb83cbb7d07a56f3ff00b3991a138e44198a16f94c7/pyzmq-26.2.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:1c17211bc037c7d88e85ed8b7d8f7e52db6dc8eca5590d162717c654550f7282", size = 1515304 }, + { url = "https://files.pythonhosted.org/packages/87/84/e8bd321aa99b72f48d4606fc5a0a920154125bd0a4608c67eab742dab087/pyzmq-26.2.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:b8f86dd868d41bea9a5f873ee13bf5551c94cf6bc51baebc6f85075971fe6eea", size = 1414712 }, + { url = "https://files.pythonhosted.org/packages/e1/bf/c67fd638c2f9fbbab8090a3ee779370b97c82b84cc12d0c498b285d7b2c0/pyzmq-26.2.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:77eb0968da535cba0470a5165468b2cac7772cfb569977cff92e240f57e31bef", size = 673129 }, + { url = "https://files.pythonhosted.org/packages/86/94/99085a3f492aa538161cbf27246e8886ff850e113e0c294a5b8245f13b52/pyzmq-26.2.0-cp311-cp311-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6ace4f71f1900a548f48407fc9be59c6ba9d9aaf658c2eea6cf2779e72f9f317", size = 910107 }, + { url = "https://files.pythonhosted.org/packages/31/1d/346809e8a9b999646d03f21096428453465b1bca5cd5c64ecd048d9ecb01/pyzmq-26.2.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:92a78853d7280bffb93df0a4a6a2498cba10ee793cc8076ef797ef2f74d107cf", size = 867960 }, + { url = "https://files.pythonhosted.org/packages/ab/68/6fb6ae5551846ad5beca295b7bca32bf0a7ce19f135cb30e55fa2314e6b6/pyzmq-26.2.0-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:689c5d781014956a4a6de61d74ba97b23547e431e9e7d64f27d4922ba96e9d6e", size = 869204 }, + { url = "https://files.pythonhosted.org/packages/0f/f9/18417771dee223ccf0f48e29adf8b4e25ba6d0e8285e33bcbce078070bc3/pyzmq-26.2.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:0aca98bc423eb7d153214b2df397c6421ba6373d3397b26c057af3c904452e37", size = 1203351 }, + { url = "https://files.pythonhosted.org/packages/e0/46/f13e67fe0d4f8a2315782cbad50493de6203ea0d744610faf4d5f5b16e90/pyzmq-26.2.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:1f3496d76b89d9429a656293744ceca4d2ac2a10ae59b84c1da9b5165f429ad3", size = 1514204 }, + { url = "https://files.pythonhosted.org/packages/50/11/ddcf7343b7b7a226e0fc7b68cbf5a5bb56291fac07f5c3023bb4c319ebb4/pyzmq-26.2.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:5c2b3bfd4b9689919db068ac6c9911f3fcb231c39f7dd30e3138be94896d18e6", size = 1414339 }, + { url = "https://files.pythonhosted.org/packages/4f/ef/5a23ec689ff36d7625b38d121ef15abfc3631a9aecb417baf7a4245e4124/pyzmq-26.2.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:55cf66647e49d4621a7e20c8d13511ef1fe1efbbccf670811864452487007e08", size = 665923 }, + { url = "https://files.pythonhosted.org/packages/ae/61/d436461a47437d63c6302c90724cf0981883ec57ceb6073873f32172d676/pyzmq-26.2.0-cp312-cp312-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4661c88db4a9e0f958c8abc2b97472e23061f0bc737f6f6179d7a27024e1faa5", size = 903400 }, + { url = "https://files.pythonhosted.org/packages/47/42/fc6d35ecefe1739a819afaf6f8e686f7f02a4dd241c78972d316f403474c/pyzmq-26.2.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ea7f69de383cb47522c9c208aec6dd17697db7875a4674c4af3f8cfdac0bdeae", size = 860034 }, + { url = "https://files.pythonhosted.org/packages/07/3b/44ea6266a6761e9eefaa37d98fabefa112328808ac41aa87b4bbb668af30/pyzmq-26.2.0-cp312-cp312-manylinux_2_28_x86_64.whl", hash = "sha256:7f98f6dfa8b8ccaf39163ce872bddacca38f6a67289116c8937a02e30bbe9711", size = 860579 }, + { url = "https://files.pythonhosted.org/packages/38/6f/4df2014ab553a6052b0e551b37da55166991510f9e1002c89cab7ce3b3f2/pyzmq-26.2.0-cp312-cp312-musllinux_1_1_aarch64.whl", hash = "sha256:e3e0210287329272539eea617830a6a28161fbbd8a3271bf4150ae3e58c5d0e6", size = 1196246 }, + { url = "https://files.pythonhosted.org/packages/38/9d/ee240fc0c9fe9817f0c9127a43238a3e28048795483c403cc10720ddef22/pyzmq-26.2.0-cp312-cp312-musllinux_1_1_i686.whl", hash = "sha256:6b274e0762c33c7471f1a7471d1a2085b1a35eba5cdc48d2ae319f28b6fc4de3", size = 1507441 }, + { url = "https://files.pythonhosted.org/packages/85/4f/01711edaa58d535eac4a26c294c617c9a01f09857c0ce191fd574d06f359/pyzmq-26.2.0-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:29c6a4635eef69d68a00321e12a7d2559fe2dfccfa8efae3ffb8e91cd0b36a8b", size = 1406498 }, + { url = "https://files.pythonhosted.org/packages/68/ba/f4280c58ff71f321602a6e24fd19879b7e79793fb8ab14027027c0fb58ef/pyzmq-26.2.0-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4d80b1dd99c1942f74ed608ddb38b181b87476c6a966a88a950c7dee118fdf50", size = 665485 }, + { url = "https://files.pythonhosted.org/packages/77/b5/c987a5c53c7d8704216f29fc3d810b32f156bcea488a940e330e1bcbb88d/pyzmq-26.2.0-cp313-cp313-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8c997098cc65e3208eca09303630e84d42718620e83b733d0fd69543a9cab9cb", size = 903484 }, + { url = "https://files.pythonhosted.org/packages/29/c9/07da157d2db18c72a7eccef8e684cefc155b712a88e3d479d930aa9eceba/pyzmq-26.2.0-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7ad1bc8d1b7a18497dda9600b12dc193c577beb391beae5cd2349184db40f187", size = 859981 }, + { url = "https://files.pythonhosted.org/packages/43/09/e12501bd0b8394b7d02c41efd35c537a1988da67fc9c745cae9c6c776d31/pyzmq-26.2.0-cp313-cp313-manylinux_2_28_x86_64.whl", hash = "sha256:bea2acdd8ea4275e1278350ced63da0b166421928276c7c8e3f9729d7402a57b", size = 860334 }, + { url = "https://files.pythonhosted.org/packages/eb/ff/f5ec1d455f8f7385cc0a8b2acd8c807d7fade875c14c44b85c1bddabae21/pyzmq-26.2.0-cp313-cp313-musllinux_1_1_aarch64.whl", hash = "sha256:23f4aad749d13698f3f7b64aad34f5fc02d6f20f05999eebc96b89b01262fb18", size = 1196179 }, + { url = "https://files.pythonhosted.org/packages/ec/8a/bb2ac43295b1950fe436a81fc5b298be0b96ac76fb029b514d3ed58f7b27/pyzmq-26.2.0-cp313-cp313-musllinux_1_1_i686.whl", hash = "sha256:a4f96f0d88accc3dbe4a9025f785ba830f968e21e3e2c6321ccdfc9aef755115", size = 1507668 }, + { url = "https://files.pythonhosted.org/packages/a9/49/dbc284ebcfd2dca23f6349227ff1616a7ee2c4a35fe0a5d6c3deff2b4fed/pyzmq-26.2.0-cp313-cp313-musllinux_1_1_x86_64.whl", hash = "sha256:ced65e5a985398827cc9276b93ef6dfabe0273c23de8c7931339d7e141c2818e", size = 1406539 }, + { url = "https://files.pythonhosted.org/packages/82/86/3fe917870e15ee1c3ad48229a2a64458e36036e64b4afa9659045d82bfa8/pyzmq-26.2.0-cp313-cp313t-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:794a4562dcb374f7dbbfb3f51d28fb40123b5a2abadee7b4091f93054909add5", size = 653242 }, + { url = "https://files.pythonhosted.org/packages/50/2d/242e7e6ef6c8c19e6cb52d095834508cd581ffb925699fd3c640cdc758f1/pyzmq-26.2.0-cp313-cp313t-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:aee22939bb6075e7afededabad1a56a905da0b3c4e3e0c45e75810ebe3a52672", size = 888404 }, + { url = "https://files.pythonhosted.org/packages/ac/11/7270566e1f31e4ea73c81ec821a4b1688fd551009a3d2bab11ec66cb1e8f/pyzmq-26.2.0-cp313-cp313t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2ae90ff9dad33a1cfe947d2c40cb9cb5e600d759ac4f0fd22616ce6540f72797", size = 845858 }, + { url = "https://files.pythonhosted.org/packages/91/d5/72b38fbc69867795c8711bdd735312f9fef1e3d9204e2f63ab57085434b9/pyzmq-26.2.0-cp313-cp313t-manylinux_2_28_x86_64.whl", hash = "sha256:43a47408ac52647dfabbc66a25b05b6a61700b5165807e3fbd40063fcaf46386", size = 847375 }, + { url = "https://files.pythonhosted.org/packages/dd/9a/10ed3c7f72b4c24e719c59359fbadd1a27556a28b36cdf1cd9e4fb7845d5/pyzmq-26.2.0-cp313-cp313t-musllinux_1_1_aarch64.whl", hash = "sha256:25bf2374a2a8433633c65ccb9553350d5e17e60c8eb4de4d92cc6bd60f01d306", size = 1183489 }, + { url = "https://files.pythonhosted.org/packages/72/2d/8660892543fabf1fe41861efa222455811adac9f3c0818d6c3170a1153e3/pyzmq-26.2.0-cp313-cp313t-musllinux_1_1_i686.whl", hash = "sha256:007137c9ac9ad5ea21e6ad97d3489af654381324d5d3ba614c323f60dab8fae6", size = 1492932 }, + { url = "https://files.pythonhosted.org/packages/7b/d6/32fd69744afb53995619bc5effa2a405ae0d343cd3e747d0fbc43fe894ee/pyzmq-26.2.0-cp313-cp313t-musllinux_1_1_x86_64.whl", hash = "sha256:470d4a4f6d48fb34e92d768b4e8a5cc3780db0d69107abf1cd7ff734b9766eb0", size = 1392485 }, + { url = "https://files.pythonhosted.org/packages/ed/69/0529b59ac667ea8bfe8796ac71796b688fbb42ff78e06525dabfed3bc7ae/pyzmq-26.2.0-cp39-cp39-manylinux_2_12_i686.manylinux2010_i686.whl", hash = "sha256:d049df610ac811dcffdc147153b414147428567fbbc8be43bb8885f04db39d98", size = 908009 }, + { url = "https://files.pythonhosted.org/packages/6e/bd/3ff3e1172f12f55769793a3a334e956ec2886805ebfb2f64756b6b5c6a1a/pyzmq-26.2.0-cp39-cp39-manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:05590cdbc6b902101d0e65d6a4780af14dc22914cc6ab995d99b85af45362cc9", size = 862078 }, + { url = "https://files.pythonhosted.org/packages/c3/ec/ab13585c3a1f48e2874253844c47b194d56eb25c94718691349c646f336f/pyzmq-26.2.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c811cfcd6a9bf680236c40c6f617187515269ab2912f3d7e8c0174898e2519db", size = 673756 }, + { url = "https://files.pythonhosted.org/packages/1e/be/febcd4b04dd50ee6d514dfbc33a3d5d9cb38ec9516e02bbfc929baa0f141/pyzmq-26.2.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:6835dd60355593de10350394242b5757fbbd88b25287314316f266e24c61d073", size = 1203684 }, + { url = "https://files.pythonhosted.org/packages/16/28/304150e71afd2df3b82f52f66c0d8ab9ac6fe1f1ffdf92bad4c8cc91d557/pyzmq-26.2.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:bc6bee759a6bddea5db78d7dcd609397449cb2d2d6587f48f3ca613b19410cfc", size = 1515864 }, + { url = "https://files.pythonhosted.org/packages/18/89/8d48d8cd505c12a1f5edee597cc32ffcedc65fd8d2603aebaaedc38a7041/pyzmq-26.2.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:c530e1eecd036ecc83c3407f77bb86feb79916d4a33d11394b8234f3bd35b940", size = 1415383 }, + { url = "https://files.pythonhosted.org/packages/77/8f/6ce54f8979a01656e894946db6299e2273fcee21c8e5fa57c6295ef11f57/pyzmq-26.2.0-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8b435f2753621cd36e7c1762156815e21c985c72b19135dac43a7f4f31d28dd1", size = 565701 }, + { url = "https://files.pythonhosted.org/packages/ee/1c/bf8cd66730a866b16db8483286078892b7f6536f8c389fb46e4beba0a970/pyzmq-26.2.0-pp310-pypy310_pp73-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:160c7e0a5eb178011e72892f99f918c04a131f36056d10d9c1afb223fc952c2d", size = 794312 }, + { url = "https://files.pythonhosted.org/packages/71/43/91fa4ff25bbfdc914ab6bafa0f03241d69370ef31a761d16bb859f346582/pyzmq-26.2.0-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2c4a71d5d6e7b28a47a394c0471b7e77a0661e2d651e7ae91e0cab0a587859ca", size = 752775 }, + { url = "https://files.pythonhosted.org/packages/da/f2/8054574d77c269c31d055d4daf3d8407adf61ea384a50c8d14b158551d09/pyzmq-26.2.0-pp39-pypy39_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:35cffef589bcdc587d06f9149f8d5e9e8859920a071df5a2671de2213bef592a", size = 565698 }, + { url = "https://files.pythonhosted.org/packages/77/21/c3ad93236d1d60eea10b67528f55e7db115a9d32e2bf163fcf601f85e9cc/pyzmq-26.2.0-pp39-pypy39_pp73-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:18c8dc3b7468d8b4bdf60ce9d7141897da103c7a4690157b32b60acb45e333e6", size = 794307 }, + { url = "https://files.pythonhosted.org/packages/6a/49/e95b491724500fcb760178ce8db39b923429e328e57bcf9162e32c2c187c/pyzmq-26.2.0-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7133d0a1677aec369d67dd78520d3fa96dd7f3dcec99d66c1762870e5ea1a50a", size = 752769 }, + { url = "https://files.pythonhosted.org/packages/9b/a9/50c9c06762b30792f71aaad8d1886748d39c4bffedc1171fbc6ad2b92d67/pyzmq-26.2.0-pp39-pypy39_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:6a96179a24b14fa6428cbfc08641c779a53f8fcec43644030328f44034c7f1f4", size = 751338 }, +] + [[package]] name = "regex" version = "2024.9.11" @@ -1091,22 +2031,22 @@ jupyter = [ [[package]] name = "ruff" -version = "0.8.3" +version = "0.8.4" source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/bf/5e/683c7ef7a696923223e7d95ca06755d6e2acbc5fd8382b2912a28008137c/ruff-0.8.3.tar.gz", hash = "sha256:5e7558304353b84279042fc584a4f4cb8a07ae79b2bf3da1a7551d960b5626d3", size = 3378522 } +sdist = { url = "https://files.pythonhosted.org/packages/34/37/9c02181ef38d55b77d97c68b78e705fd14c0de0e5d085202bb2b52ce5be9/ruff-0.8.4.tar.gz", hash = "sha256:0d5f89f254836799af1615798caa5f80b7f935d7a670fad66c5007928e57ace8", size = 3402103 } wheels = [ - { url = "https://files.pythonhosted.org/packages/f8/c4/bfdbb8b9c419ff3b52479af8581026eeaac3764946fdb463dec043441b7d/ruff-0.8.3-py3-none-linux_armv6l.whl", hash = "sha256:8d5d273ffffff0acd3db5bf626d4b131aa5a5ada1276126231c4174543ce20d6", size = 10535860 }, - { url = "https://files.pythonhosted.org/packages/91/5a/642ed8f1ba23ffc2dd347697e01eef3c42fad6ac76603be4a8c3a9d6311e/ruff-0.8.3-py3-none-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9c0a60a825e3e177116c84009d5ebaa90cf40dfab56e1358d1df4e29a9a14b13", size = 10797597 }, - { url = "https://files.pythonhosted.org/packages/30/25/2e654bc7226da09a49730a1a2ea6e89f843b362db80b4b2a7a4f948ac986/ruff-0.8.3-py3-none-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:75fb782f4db39501210ac093c79c3de581d306624575eddd7e4e13747e61ba18", size = 10307244 }, - { url = "https://files.pythonhosted.org/packages/c0/2d/a224d56bcd4383583db53c2b8f410ebf1200866984aa6eb9b5a70f04e71f/ruff-0.8.3-py3-none-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7f26bc76a133ecb09a38b7868737eded6941b70a6d34ef53a4027e83913b6502", size = 11362439 }, - { url = "https://files.pythonhosted.org/packages/82/01/03e2857f9c371b8767d3e909f06a33bbdac880df17f17f93d6f6951c3381/ruff-0.8.3-py3-none-manylinux_2_17_ppc64.manylinux2014_ppc64.whl", hash = "sha256:01b14b2f72a37390c1b13477c1c02d53184f728be2f3ffc3ace5b44e9e87b90d", size = 12078538 }, - { url = "https://files.pythonhosted.org/packages/af/ae/ff7f97b355da16d748ceec50e1604a8215d3659b36b38025a922e0612e9b/ruff-0.8.3-py3-none-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:53babd6e63e31f4e96ec95ea0d962298f9f0d9cc5990a1bbb023a6baf2503a82", size = 11616172 }, - { url = "https://files.pythonhosted.org/packages/6a/d0/6156d4d1e53ebd17747049afe801c5d7e3014d9b2f398b9236fe36ba4320/ruff-0.8.3-py3-none-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:1ae441ce4cf925b7f363d33cd6570c51435972d697e3e58928973994e56e1452", size = 12919886 }, - { url = "https://files.pythonhosted.org/packages/4e/84/affcb30bacb94f6036a128ad5de0e29f543d3f67ee42b490b17d68e44b8a/ruff-0.8.3-py3-none-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d7c65bc0cadce32255e93c57d57ecc2cca23149edd52714c0c5d6fa11ec328cd", size = 11212599 }, - { url = "https://files.pythonhosted.org/packages/60/b9/5694716bdefd8f73df7c0104334156c38fb0f77673d2966a5a1345bab94d/ruff-0.8.3-py3-none-musllinux_1_2_aarch64.whl", hash = "sha256:5be450bb18f23f0edc5a4e5585c17a56ba88920d598f04a06bd9fd76d324cb20", size = 10784637 }, - { url = "https://files.pythonhosted.org/packages/24/7e/0e8f835103ac7da81c3663eedf79dec8359e9ae9a3b0d704bae50be59176/ruff-0.8.3-py3-none-musllinux_1_2_armv7l.whl", hash = "sha256:8faeae3827eaa77f5721f09b9472a18c749139c891dbc17f45e72d8f2ca1f8fc", size = 10390591 }, - { url = "https://files.pythonhosted.org/packages/27/da/180ec771fc01c004045962ce017ca419a0281f4bfaf867ed0020f555b56e/ruff-0.8.3-py3-none-musllinux_1_2_i686.whl", hash = "sha256:db503486e1cf074b9808403991663e4277f5c664d3fe237ee0d994d1305bb060", size = 10894298 }, - { url = "https://files.pythonhosted.org/packages/6d/f8/29f241742ed3954eb2222314b02db29f531a15cab3238d1295e8657c5f18/ruff-0.8.3-py3-none-musllinux_1_2_x86_64.whl", hash = "sha256:6567be9fb62fbd7a099209257fef4ad2c3153b60579818b31a23c886ed4147ea", size = 11275965 }, + { url = "https://files.pythonhosted.org/packages/05/67/f480bf2f2723b2e49af38ed2be75ccdb2798fca7d56279b585c8f553aaab/ruff-0.8.4-py3-none-linux_armv6l.whl", hash = "sha256:58072f0c06080276804c6a4e21a9045a706584a958e644353603d36ca1eb8a60", size = 10546415 }, + { url = "https://files.pythonhosted.org/packages/0e/f0/afa0d2191af495ac82d4cbbfd7a94e3df6f62a04ca412033e073b871fc6d/ruff-0.8.4-py3-none-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e248b1f0fa2749edd3350a2a342b67b43a2627434c059a063418e3d375cfe643", size = 10805522 }, + { url = "https://files.pythonhosted.org/packages/12/57/5d1e9a0fd0c228e663894e8e3a8e7063e5ee90f8e8e60cf2085f362bfa1a/ruff-0.8.4-py3-none-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:bf197b98ed86e417412ee3b6c893f44c8864f816451441483253d5ff22c0e81e", size = 10306763 }, + { url = "https://files.pythonhosted.org/packages/04/df/f069fdb02e408be8aac6853583572a2873f87f866fe8515de65873caf6b8/ruff-0.8.4-py3-none-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c41319b85faa3aadd4d30cb1cffdd9ac6b89704ff79f7664b853785b48eccdf3", size = 11359574 }, + { url = "https://files.pythonhosted.org/packages/d3/04/37c27494cd02e4a8315680debfc6dfabcb97e597c07cce0044db1f9dfbe2/ruff-0.8.4-py3-none-manylinux_2_17_ppc64.manylinux2014_ppc64.whl", hash = "sha256:9f8402b7c4f96463f135e936d9ab77b65711fcd5d72e5d67597b543bbb43cf3f", size = 12094851 }, + { url = "https://files.pythonhosted.org/packages/81/b1/c5d7fb68506cab9832d208d03ea4668da9a9887a4a392f4f328b1bf734ad/ruff-0.8.4-py3-none-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e4e56b3baa9c23d324ead112a4fdf20db9a3f8f29eeabff1355114dd96014604", size = 11655539 }, + { url = "https://files.pythonhosted.org/packages/ef/38/8f8f2c8898dc8a7a49bc340cf6f00226917f0f5cb489e37075bcb2ce3671/ruff-0.8.4-py3-none-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:736272574e97157f7edbbb43b1d046125fce9e7d8d583d5d65d0c9bf2c15addf", size = 12912805 }, + { url = "https://files.pythonhosted.org/packages/06/dd/fa6660c279f4eb320788876d0cff4ea18d9af7d9ed7216d7bd66877468d0/ruff-0.8.4-py3-none-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e5fe710ab6061592521f902fca7ebcb9fabd27bc7c57c764298b1c1f15fff720", size = 11205976 }, + { url = "https://files.pythonhosted.org/packages/a8/d7/de94cc89833b5de455750686c17c9e10f4e1ab7ccdc5521b8fe911d1477e/ruff-0.8.4-py3-none-musllinux_1_2_aarch64.whl", hash = "sha256:13e9ec6d6b55f6da412d59953d65d66e760d583dd3c1c72bf1f26435b5bfdbae", size = 10792039 }, + { url = "https://files.pythonhosted.org/packages/6d/15/3e4906559248bdbb74854af684314608297a05b996062c9d72e0ef7c7097/ruff-0.8.4-py3-none-musllinux_1_2_armv7l.whl", hash = "sha256:97d9aefef725348ad77d6db98b726cfdb075a40b936c7984088804dfd38268a7", size = 10400088 }, + { url = "https://files.pythonhosted.org/packages/a2/21/9ed4c0e8133cb4a87a18d470f534ad1a8a66d7bec493bcb8bda2d1a5d5be/ruff-0.8.4-py3-none-musllinux_1_2_i686.whl", hash = "sha256:ab78e33325a6f5374e04c2ab924a3367d69a0da36f8c9cb6b894a62017506111", size = 10900814 }, + { url = "https://files.pythonhosted.org/packages/0d/5d/122a65a18955bd9da2616b69bc839351f8baf23b2805b543aa2f0aed72b5/ruff-0.8.4-py3-none-musllinux_1_2_x86_64.whl", hash = "sha256:8ef06f66f4a05c3ddbc9121a8b0cecccd92c5bf3dd43b5472ffe40b8ca10f0f8", size = 11268828 }, ] [[package]] @@ -1169,6 +2109,25 @@ wheels = [ { url = "https://files.pythonhosted.org/packages/8d/02/1165905f14962174e6569076bcc3315809ae1291ed14de6448cc151eedfd/scipy-1.13.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:a014c2b3697bde71724244f63de2476925596c24285c7a637364761f8710891c", size = 38845637 }, ] +[[package]] +name = "sentencepiece" +version = "0.2.0" +source = { registry = "https://download.pytorch.org/whl/nightly/cu126" } +wheels = [ + { url = "https://download.pytorch.org/whl/nightly/sentencepiece-0.2.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/sentencepiece-0.2.0-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl" }, + { url = "https://download.pytorch.org/whl/nightly/sentencepiece-0.2.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/sentencepiece-0.2.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/sentencepiece-0.2.0-cp311-cp311-manylinux_2_17_i686.manylinux2014_i686.whl" }, + { url = "https://download.pytorch.org/whl/nightly/sentencepiece-0.2.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/sentencepiece-0.2.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/sentencepiece-0.2.0-cp312-cp312-manylinux_2_17_i686.manylinux2014_i686.whl" }, + { url = "https://download.pytorch.org/whl/nightly/sentencepiece-0.2.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/sentencepiece-0.2.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/sentencepiece-0.2.0-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl" }, + { url = "https://download.pytorch.org/whl/nightly/sentencepiece-0.2.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, +] + [[package]] name = "setuptools" version = "70.2.0" @@ -1177,6 +2136,24 @@ wheels = [ { url = "https://download.pytorch.org/whl/nightly/setuptools-70.2.0-py3-none-any.whl" }, ] +[[package]] +name = "six" +version = "1.17.0" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/94/e7/b2c673351809dca68a0e064b6af791aa332cf192da575fd474ed7d6f16a2/six-1.17.0.tar.gz", hash = "sha256:ff70335d468e7eb6ec65b95b99d3a2836546063f63acc5171de367e834932a81", size = 34031 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/b7/ce/149a00dd41f10bc29e5921b496af8b574d8413afcd5e30dfa0ed46c2cc5e/six-1.17.0-py2.py3-none-any.whl", hash = "sha256:4721f391ed90541fddacab5acf947aa0d3dc7d27b2e1e8eda2be8970586c3274", size = 11050 }, +] + +[[package]] +name = "sniffio" +version = "1.3.1" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/a2/87/a6771e1546d97e7e041b6ae58d80074f81b7d5121207425c964ddf5cfdbd/sniffio-1.3.1.tar.gz", hash = "sha256:f4324edc670a0f49750a81b895f35c3adb843cca46f0530f79fc1babb23789dc", size = 20372 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/e9/44/75a9c9421471a6c4805dbf2356f7c181a29c1879239abab1ea2cc8f38b40/sniffio-1.3.1-py3-none-any.whl", hash = "sha256:2f6da418d1f1e0fddd844478f41680e794e6051915791a034ff65e5f100525a2", size = 10235 }, +] + [[package]] name = "stack-data" version = "0.6.3" @@ -1191,6 +2168,28 @@ wheels = [ { url = "https://files.pythonhosted.org/packages/f1/7b/ce1eafaf1a76852e2ec9b22edecf1daa58175c090266e9f6c64afcd81d91/stack_data-0.6.3-py3-none-any.whl", hash = "sha256:d5558e0c25a4cb0853cddad3d77da9891a08cb85dd9f9f91b9f8cd66e511e695", size = 24521 }, ] +[[package]] +name = "starlette" +version = "0.41.3" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "anyio", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "typing-extensions", marker = "(python_full_version < '3.10' and sys_platform == 'linux') or (python_full_version < '3.10' and sys_platform == 'windows')" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/1a/4c/9b5764bd22eec91c4039ef4c55334e9187085da2d8a2df7bd570869aae18/starlette-0.41.3.tar.gz", hash = "sha256:0e4ab3d16522a255be6b28260b938eae2482f98ce5cc934cb08dce8dc3ba5835", size = 2574159 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/96/00/2b325970b3060c7cecebab6d295afe763365822b1306a12eeab198f74323/starlette-0.41.3-py3-none-any.whl", hash = "sha256:44cedb2b7c77a9de33a8b74b2b90e9f50d11fcf25d8270ea525ad71a25374ff7", size = 73225 }, +] + +[[package]] +name = "strenum" +version = "0.4.15" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/85/ad/430fb60d90e1d112a62ff57bdd1f286ec73a2a0331272febfddd21f330e1/StrEnum-0.4.15.tar.gz", hash = "sha256:878fb5ab705442070e4dd1929bb5e2249511c0bcf2b0eeacf3bcd80875c82eff", size = 23384 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/81/69/297302c5f5f59c862faa31e6cb9a4cd74721cd1e052b38e464c5b402df8b/StrEnum-0.4.15-py3-none-any.whl", hash = "sha256:a30cda4af7cc6b5bf52c8055bc4bf4b2b6b14a93b574626da33df53cf7740659", size = 8851 }, +] + [[package]] name = "sympy" version = "1.13.1" @@ -1202,6 +2201,15 @@ wheels = [ { url = "https://download.pytorch.org/whl/nightly/sympy-1.13.1-py3-none-any.whl" }, ] +[[package]] +name = "tensorrt" +version = "10.7.0" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "tensorrt-cu12", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/76/a6/791f1e15e708be8c9d958236c1659d1d27fe8a31fc84f011f9920c446d7a/tensorrt-10.7.0.tar.gz", hash = "sha256:2c74ee72c450ef055e894e04f27510094f81e575a284af9d4b4b06a4369d0e2d", size = 16401 } + [[package]] name = "tensorrt-cu12" version = "10.7.0" @@ -1232,25 +2240,113 @@ dependencies = [ ] sdist = { url = "https://files.pythonhosted.org/packages/e7/72/69dd1a6ae202c826f61dcd5bc3019449ed1c961d0b60804311d2493403c8/tensorrt_cu12_libs-10.7.0.tar.gz", hash = "sha256:45f6419adf59ad6532d44f945f1efe79a2975089bf7f92a4374945df40951c02", size = 699 } +[[package]] +name = "tensorrt-llm" +version = "0.16.0" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "accelerate", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "aenum", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "build", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "click", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "click-option-group", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "colored", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "cuda-python", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "diffusers", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "evaluate", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "fastapi", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "h5py", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "httpx", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "lark", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "mpi4py", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "mpmath", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "numpy", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "nvidia-modelopt", extra = ["torch"], marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "nvidia-nccl-cu12", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "onnx", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "onnx-graphsurgeon", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "openai", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "optimum", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "ordered-set", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "pandas", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "pillow", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "polygraphy", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "psutil", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "pulp", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "pydantic", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "pynvml", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "pyzmq", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "sentencepiece", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "setuptools", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "strenum", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "tensorrt", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "torch", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "torchvision", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "transformers", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "uvicorn", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "wheel", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/de/66/91b2582f56f370b54be501bd9fb8e4b90bdc16fe400173b6aa04a2ab7c70/tensorrt_llm-0.16.0.tar.gz", hash = "sha256:f0337f979b85f2579ad6fd3d3062cb96b1ccc3876e635ca8070617572ec12008", size = 1240 } + [[package]] name = "tokenizers" -version = "0.21.0" +version = "0.20.3" source = { registry = "https://pypi.org/simple" } dependencies = [ { name = "huggingface-hub", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, ] -sdist = { url = "https://files.pythonhosted.org/packages/20/41/c2be10975ca37f6ec40d7abd7e98a5213bb04f284b869c1a24e6504fd94d/tokenizers-0.21.0.tar.gz", hash = "sha256:ee0894bf311b75b0c03079f33859ae4b2334d675d4e93f5a4132e1eae2834fe4", size = 343021 } -wheels = [ - { url = "https://files.pythonhosted.org/packages/f7/14/83429177c19364df27d22bc096d4c2e431e0ba43e56c525434f1f9b0fd00/tokenizers-0.21.0-cp39-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6b177fb54c4702ef611de0c069d9169f0004233890e0c4c5bd5508ae05abf193", size = 2903304 }, - { url = "https://files.pythonhosted.org/packages/7e/db/3433eab42347e0dc5452d8fcc8da03f638c9accffefe5a7c78146666964a/tokenizers-0.21.0-cp39-abi3-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:6b43779a269f4629bebb114e19c3fca0223296ae9fea8bb9a7a6c6fb0657ff8e", size = 2804378 }, - { url = "https://files.pythonhosted.org/packages/57/8b/7da5e6f89736c2ade02816b4733983fca1c226b0c42980b1ae9dc8fcf5cc/tokenizers-0.21.0-cp39-abi3-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:9aeb255802be90acfd363626753fda0064a8df06031012fe7d52fd9a905eb00e", size = 3095488 }, - { url = "https://files.pythonhosted.org/packages/4d/f6/5ed6711093dc2c04a4e03f6461798b12669bc5a17c8be7cce1240e0b5ce8/tokenizers-0.21.0-cp39-abi3-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d8b09dbeb7a8d73ee204a70f94fc06ea0f17dcf0844f16102b9f414f0b7463ba", size = 3121410 }, - { url = "https://files.pythonhosted.org/packages/81/42/07600892d48950c5e80505b81411044a2d969368cdc0d929b1c847bf6697/tokenizers-0.21.0-cp39-abi3-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:400832c0904f77ce87c40f1a8a27493071282f785724ae62144324f171377273", size = 3388821 }, - { url = "https://files.pythonhosted.org/packages/22/06/69d7ce374747edaf1695a4f61b83570d91cc8bbfc51ccfecf76f56ab4aac/tokenizers-0.21.0-cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e84ca973b3a96894d1707e189c14a774b701596d579ffc7e69debfc036a61a04", size = 3008868 }, - { url = "https://files.pythonhosted.org/packages/c8/69/54a0aee4d576045b49a0eb8bffdc495634309c823bf886042e6f46b80058/tokenizers-0.21.0-cp39-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:eb7202d231b273c34ec67767378cd04c767e967fda12d4a9e36208a34e2f137e", size = 8975831 }, - { url = "https://files.pythonhosted.org/packages/f7/f3/b776061e4f3ebf2905ba1a25d90380aafd10c02d406437a8ba22d1724d76/tokenizers-0.21.0-cp39-abi3-musllinux_1_2_armv7l.whl", hash = "sha256:089d56db6782a73a27fd8abf3ba21779f5b85d4a9f35e3b493c7bbcbbf0d539b", size = 8920746 }, - { url = "https://files.pythonhosted.org/packages/d8/ee/ce83d5ec8b6844ad4c3ecfe3333d58ecc1adc61f0878b323a15355bcab24/tokenizers-0.21.0-cp39-abi3-musllinux_1_2_i686.whl", hash = "sha256:c87ca3dc48b9b1222d984b6b7490355a6fdb411a2d810f6f05977258400ddb74", size = 9161814 }, - { url = "https://files.pythonhosted.org/packages/18/07/3e88e65c0ed28fa93aa0c4d264988428eef3df2764c3126dc83e243cb36f/tokenizers-0.21.0-cp39-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:4145505a973116f91bc3ac45988a92e618a6f83eb458f49ea0790df94ee243ff", size = 9357138 }, +sdist = { url = "https://files.pythonhosted.org/packages/da/25/b1681c1c30ea3ea6e584ae3fffd552430b12faa599b558c4c4783f56d7ff/tokenizers-0.20.3.tar.gz", hash = "sha256:2278b34c5d0dd78e087e1ca7f9b1dcbf129d80211afa645f214bd6e051037539", size = 340513 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/e3/e8/0e9f81a09ab79f409eabfd99391ca519e315496694671bebca24c3e90448/tokenizers-0.20.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f128d5da1202b78fa0a10d8d938610472487da01b57098d48f7e944384362514", size = 2892896 }, + { url = "https://files.pythonhosted.org/packages/b0/72/15fdbc149e05005e99431ecd471807db2241983deafe1e704020f608f40e/tokenizers-0.20.3-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:79c4121a2e9433ad7ef0769b9ca1f7dd7fa4c0cd501763d0a030afcbc6384481", size = 2802785 }, + { url = "https://files.pythonhosted.org/packages/26/44/1f8aea48f9bb117d966b7272484671b33a509f6217a8e8544d79442c90db/tokenizers-0.20.3-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b7850fde24197fe5cd6556e2fdba53a6d3bae67c531ea33a3d7c420b90904141", size = 3086060 }, + { url = "https://files.pythonhosted.org/packages/2e/83/82ba40da99870b3a0b801cffaf4f099f088a84c7e07d32cc6ca751ce08e6/tokenizers-0.20.3-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:b357970c095dc134978a68c67d845a1e3803ab7c4fbb39195bde914e7e13cf8b", size = 3096760 }, + { url = "https://files.pythonhosted.org/packages/f3/46/7a025404201d937f86548928616c0a164308aa3998e546efdf798bf5ee9c/tokenizers-0.20.3-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a333d878c4970b72d6c07848b90c05f6b045cf9273fc2bc04a27211721ad6118", size = 3380165 }, + { url = "https://files.pythonhosted.org/packages/aa/49/15fae66ac62e49255eeedbb7f4127564b2c3f3aef2009913f525732d1a08/tokenizers-0.20.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1fd9fee817f655a8f50049f685e224828abfadd436b8ff67979fc1d054b435f1", size = 2994038 }, + { url = "https://files.pythonhosted.org/packages/f4/64/693afc9ba2393c2eed85c02bacb44762f06a29f0d1a5591fa5b40b39c0a2/tokenizers-0.20.3-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:9e7816808b402129393a435ea2a509679b41246175d6e5e9f25b8692bfaa272b", size = 8977285 }, + { url = "https://files.pythonhosted.org/packages/be/7e/6126c18694310fe07970717929e889898767c41fbdd95b9078e8aec0f9ef/tokenizers-0.20.3-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:ba96367db9d8a730d3a1d5996b4b7babb846c3994b8ef14008cd8660f55db59d", size = 9294890 }, + { url = "https://files.pythonhosted.org/packages/46/54/033b5b2ba0c3ae01e026c6f7ced147d41a2fa1c573d00a66cb97f6d7f9b3/tokenizers-0.20.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ef820880d5e4e8484e2fa54ff8d297bb32519eaa7815694dc835ace9130a3eea", size = 2892476 }, + { url = "https://files.pythonhosted.org/packages/e6/b0/cc369fb3297d61f3311cab523d16d48c869dc2f0ba32985dbf03ff811041/tokenizers-0.20.3-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:67ef4dcb8841a4988cd00dd288fb95dfc8e22ed021f01f37348fd51c2b055ba9", size = 2802775 }, + { url = "https://files.pythonhosted.org/packages/1a/74/62ad983e8ea6a63e04ed9c5be0b605056bf8aac2f0125f9b5e0b3e2b89fa/tokenizers-0.20.3-cp311-cp311-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ff1ef8bd47a02b0dc191688ccb4da53600df5d4c9a05a4b68e1e3de4823e78eb", size = 3086138 }, + { url = "https://files.pythonhosted.org/packages/6b/ac/4637ba619db25094998523f9e6f5b456e1db1f8faa770a3d925d436db0c3/tokenizers-0.20.3-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:444d188186eab3148baf0615b522461b41b1f0cd58cd57b862ec94b6ac9780f1", size = 3098076 }, + { url = "https://files.pythonhosted.org/packages/58/ce/9793f2dc2ce529369807c9c74e42722b05034af411d60f5730b720388c7d/tokenizers-0.20.3-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:37c04c032c1442740b2c2d925f1857885c07619224a533123ac7ea71ca5713da", size = 3379650 }, + { url = "https://files.pythonhosted.org/packages/50/f6/2841de926bc4118af996eaf0bdf0ea5b012245044766ffc0347e6c968e63/tokenizers-0.20.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:453c7769d22231960ee0e883d1005c93c68015025a5e4ae56275406d94a3c907", size = 2994005 }, + { url = "https://files.pythonhosted.org/packages/a3/b2/00915c4fed08e9505d37cf6eaab45b12b4bff8f6719d459abcb9ead86a4b/tokenizers-0.20.3-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:4bb31f7b2847e439766aaa9cc7bccf7ac7088052deccdb2275c952d96f691c6a", size = 8977488 }, + { url = "https://files.pythonhosted.org/packages/e9/ac/1c069e7808181ff57bcf2d39e9b6fbee9133a55410e6ebdaa89f67c32e83/tokenizers-0.20.3-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:843729bf0f991b29655a069a2ff58a4c24375a553c70955e15e37a90dd4e045c", size = 9294935 }, + { url = "https://files.pythonhosted.org/packages/27/37/d108df55daf4f0fcf1f58554692ff71687c273d870a34693066f0847be96/tokenizers-0.20.3-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:abe4e08c7d0cd6154c795deb5bf81d2122f36daf075e0c12a8b050d824ef0a64", size = 2898389 }, + { url = "https://files.pythonhosted.org/packages/b2/27/32f29da16d28f59472fa7fb38e7782069748c7e9ab9854522db20341624c/tokenizers-0.20.3-cp312-cp312-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:ca94fc1b73b3883c98f0c88c77700b13d55b49f1071dfd57df2b06f3ff7afd64", size = 2795866 }, + { url = "https://files.pythonhosted.org/packages/29/4e/8a9a3c89e128c4a40f247b501c10279d2d7ade685953407c4d94c8c0f7a7/tokenizers-0.20.3-cp312-cp312-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ef279c7e239f95c8bdd6ff319d9870f30f0d24915b04895f55b1adcf96d6c60d", size = 3085446 }, + { url = "https://files.pythonhosted.org/packages/b4/3b/a2a7962c496ebcd95860ca99e423254f760f382cd4bd376f8895783afaf5/tokenizers-0.20.3-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:16384073973f6ccbde9852157a4fdfe632bb65208139c9d0c0bd0176a71fd67f", size = 3094378 }, + { url = "https://files.pythonhosted.org/packages/1f/f4/a8a33f0192a1629a3bd0afcad17d4d221bbf9276da4b95d226364208d5eb/tokenizers-0.20.3-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:312d522caeb8a1a42ebdec87118d99b22667782b67898a76c963c058a7e41d4f", size = 3385755 }, + { url = "https://files.pythonhosted.org/packages/9e/65/c83cb3545a65a9eaa2e13b22c93d5e00bd7624b354a44adbdc93d5d9bd91/tokenizers-0.20.3-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f2b7cb962564785a83dafbba0144ecb7f579f1d57d8c406cdaa7f32fe32f18ad", size = 2997679 }, + { url = "https://files.pythonhosted.org/packages/55/e9/a80d4e592307688a67c7c59ab77e03687b6a8bd92eb5db763a2c80f93f57/tokenizers-0.20.3-cp312-cp312-musllinux_1_1_aarch64.whl", hash = "sha256:124c5882ebb88dadae1fc788a582299fcd3a8bd84fc3e260b9918cf28b8751f5", size = 8989296 }, + { url = "https://files.pythonhosted.org/packages/90/af/60c957af8d2244321124e893828f1a4817cde1a2d08d09d423b73f19bd2f/tokenizers-0.20.3-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:2b6e54e71f84c4202111a489879005cb14b92616a87417f6c102c833af961ea2", size = 9303621 }, + { url = "https://files.pythonhosted.org/packages/7c/cf/5309c2d173a6a67f9ec8697d8e710ea32418de6fd8541778032c202a1c3e/tokenizers-0.20.3-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0e3d80d89b068bc30034034b5319218c7c0a91b00af19679833f55f3becb6945", size = 2897745 }, + { url = "https://files.pythonhosted.org/packages/2c/e5/af3078e32f225e680e69d61f78855880edb8d53f5850a1834d519b2b103f/tokenizers-0.20.3-cp313-cp313-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:174a54910bed1b089226512b4458ea60d6d6fd93060254734d3bc3540953c51c", size = 2794385 }, + { url = "https://files.pythonhosted.org/packages/0b/a7/bc421fe46650cc4eb4a913a236b88c243204f32c7480684d2f138925899e/tokenizers-0.20.3-cp313-cp313-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:098b8a632b8656aa5802c46689462c5c48f02510f24029d71c208ec2c822e771", size = 3084580 }, + { url = "https://files.pythonhosted.org/packages/c6/22/97e1e95ee81f75922c9f569c23cb2b1fdc7f5a7a29c4c9fae17e63f751a6/tokenizers-0.20.3-cp313-cp313-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:78c8c143e3ae41e718588281eb3e212c2b31623c9d6d40410ec464d7d6221fb5", size = 3093581 }, + { url = "https://files.pythonhosted.org/packages/d5/14/f0df0ee3b9e516121e23c0099bccd7b9f086ba9150021a750e99b16ce56f/tokenizers-0.20.3-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:2b26b0aadb18cd8701077362ba359a06683662d5cafe3e8e8aba10eb05c037f1", size = 3385934 }, + { url = "https://files.pythonhosted.org/packages/66/52/7a171bd4929e3ffe61a29b4340fe5b73484709f92a8162a18946e124c34c/tokenizers-0.20.3-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:07d7851a72717321022f3774e84aa9d595a041d643fafa2e87fbc9b18711dac0", size = 2997311 }, + { url = "https://files.pythonhosted.org/packages/7c/64/f1993bb8ebf775d56875ca0d50a50f2648bfbbb143da92fe2e6ceeb4abd5/tokenizers-0.20.3-cp313-cp313-musllinux_1_1_aarch64.whl", hash = "sha256:bd44e48a430ada902c6266a8245f5036c4fe744fcb51f699999fbe82aa438797", size = 8988601 }, + { url = "https://files.pythonhosted.org/packages/d6/3f/49fa63422159bbc2f2a4ac5bfc597d04d4ec0ad3d2ef46649b5e9a340e37/tokenizers-0.20.3-cp313-cp313-musllinux_1_1_x86_64.whl", hash = "sha256:a4c186bb006ccbe1f5cc4e0380d1ce7806f5955c244074fd96abc55e27b77f01", size = 9303950 }, + { url = "https://files.pythonhosted.org/packages/b0/39/073836c1d73e63268b1c67a682a8ba23e2688a43e737166be45ab8243701/tokenizers-0.20.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7eb2fb1c432f5746b22f8a7f09fc18c4156cb0031c77f53cb19379d82d43297a", size = 2893676 }, + { url = "https://files.pythonhosted.org/packages/c1/d9/b9ff819c3df4bc73ad93629804f7b85321a78bc2da4f54fb774a90e995c6/tokenizers-0.20.3-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:bfa8d029bb156181b006643309d6b673615a24e4ed24cf03aa191d599b996f51", size = 2804173 }, + { url = "https://files.pythonhosted.org/packages/3e/d5/6b2b519ba2d9a6d3435f22918f0ad5850c40cf5357f6d989e6d68ef40fb9/tokenizers-0.20.3-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6f90549622de3bf476ad9f1dd6f3f952ec3ed6ab8615ae88ef060d0c5bfad55d", size = 3086866 }, + { url = "https://files.pythonhosted.org/packages/01/e1/d96e90ef872dd9b3a4b7a78874411f1c48476019f95a87a2cfd54c470a57/tokenizers-0.20.3-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:a1d469c74eebf5c43fd61cd9b030e271d17198edd7bd45392e03a3c091d7d6d4", size = 3099004 }, + { url = "https://files.pythonhosted.org/packages/0c/6a/a94248dc5915907e18d55c9739cd018f5aeb4146f198622f45f9748dcb9f/tokenizers-0.20.3-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:bee8f53b2594749f4460d53253bae55d718f04e9b633efa0f5df8938bd98e4f0", size = 3381574 }, + { url = "https://files.pythonhosted.org/packages/29/9e/c95f8821d6bc93eba7c5db95e6299c009db523d1c646da8563b42ad892c4/tokenizers-0.20.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:938441babf3e5720e4459e306ef2809fb267680df9d1ff2873458b22aef60248", size = 2994953 }, + { url = "https://files.pythonhosted.org/packages/95/ff/01fdcf9a77776730baf63a9f66adf75c3aa4bdb1bdc77c7d1a3e03b2a25e/tokenizers-0.20.3-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:7310ab23d7b0caebecc0e8be11a1146f320f5f07284000f6ea54793e83de1b75", size = 8977698 }, + { url = "https://files.pythonhosted.org/packages/ef/2d/8b823741c64e9726b82076fa09f6d66285b61bd2c77e109871415b1ed9e2/tokenizers-0.20.3-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:16121eb030a2b13094cfec936b0c12e8b4063c5f839591ea7d0212336d8f9921", size = 9295649 }, + { url = "https://files.pythonhosted.org/packages/8e/c1/6af62ef61316f33ecf785bbb2bee4292f34ea62b491d4480ad9b09acf6b6/tokenizers-0.20.3-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:39270a7050deaf50f7caff4c532c01b3c48f6608d42b3eacdebdc6795478c8df", size = 2897936 }, + { url = "https://files.pythonhosted.org/packages/9a/0b/c076b2ff3ee6dc70c805181fbe325668b89cfee856f8dfa24cc9aa293c84/tokenizers-0.20.3-pp310-pypy310_pp73-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e005466632b1c5d2d2120f6de8aa768cc9d36cd1ab7d51d0c27a114c91a1e6ee", size = 3082688 }, + { url = "https://files.pythonhosted.org/packages/0a/60/56510124933136c2e90879e1c81603cfa753ae5a87830e3ef95056b20d8f/tokenizers-0.20.3-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a07962340b36189b6c8feda552ea1bfeee6cf067ff922a1d7760662c2ee229e5", size = 2998924 }, + { url = "https://files.pythonhosted.org/packages/68/60/4107b618b7b9155cb34ad2e0fc90946b7e71f041b642122fb6314f660688/tokenizers-0.20.3-pp310-pypy310_pp73-musllinux_1_1_aarch64.whl", hash = "sha256:55046ad3dd5f2b3c67501fcc8c9cbe3e901d8355f08a3b745e9b57894855f85b", size = 8989514 }, + { url = "https://files.pythonhosted.org/packages/e8/bd/48475818e614b73316baf37ac1e4e51b578bbdf58651812d7e55f43b88d8/tokenizers-0.20.3-pp310-pypy310_pp73-musllinux_1_1_x86_64.whl", hash = "sha256:efcf0eb939988b627558aaf2b9dc3e56d759cad2e0cfa04fcab378e4b48fc4fd", size = 9303476 }, + { url = "https://files.pythonhosted.org/packages/ce/32/37ff2ced2c169c2e7586fcd51314f59d02c60fd2eeafea527c2f9d1bb512/tokenizers-0.20.3-pp39-pypy39_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a292392f24ab9abac5cfa8197e5a6208f2e43723420217e1ceba0b4ec77816ac", size = 2897613 }, + { url = "https://files.pythonhosted.org/packages/79/e4/fdd7ad2aedaa4a3f148aa28670bf0b0856211a3fec3e6554ed6ceec9a928/tokenizers-0.20.3-pp39-pypy39_pp73-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8dcd91f4e60f62b20d83a87a84fe062035a1e3ff49a8c2bbdeb2d441c8e311f4", size = 3085434 }, + { url = "https://files.pythonhosted.org/packages/e0/b8/479ab7349faf1da001b861ea521055ad18a34a9b1053079e0c9b5c476f50/tokenizers-0.20.3-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:900991a2b8ee35961b1095db7e265342e0e42a84c1a594823d5ee9f8fb791958", size = 2998651 }, + { url = "https://files.pythonhosted.org/packages/6b/7f/3a1d5ded5f841764d67aa4c6e2e4b40d9dac5fbd2df135bccc58284a6917/tokenizers-0.20.3-pp39-pypy39_pp73-musllinux_1_1_aarch64.whl", hash = "sha256:5a8d8261ca2133d4f98aa9627c748189502b3787537ba3d7e2beb4f7cfc5d627", size = 8989010 }, + { url = "https://files.pythonhosted.org/packages/2b/a7/e0b5d5fea8cb69afdbab3c0e0cc3a02b5dd888ce0f933312f7c0ca6b017e/tokenizers-0.20.3-pp39-pypy39_pp73-musllinux_1_1_x86_64.whl", hash = "sha256:c4fd4d71e6deb6ddf99d8d0eab87d1d16f635898906e631914a9bae8ae9f2cfb", size = 9303287 }, ] [[package]] @@ -1282,7 +2378,7 @@ wheels = [ [[package]] name = "torch" -version = "2.6.0.dev20241212+cu126" +version = "2.6.0.dev20241224+cu126" source = { registry = "https://download.pytorch.org/whl/nightly/cu126" } dependencies = [ { name = "filelock", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, @@ -1302,28 +2398,28 @@ dependencies = [ { name = "nvidia-nccl-cu12", marker = "(platform_machine == 'x86_64' and platform_system == 'Linux' and sys_platform == 'linux') or (platform_machine == 'x86_64' and platform_system == 'Linux' and sys_platform == 'windows')" }, { name = "nvidia-nvjitlink-cu12", marker = "(platform_machine == 'x86_64' and platform_system == 'Linux' and sys_platform == 'linux') or (platform_machine == 'x86_64' and platform_system == 'Linux' and sys_platform == 'windows')" }, { name = "nvidia-nvtx-cu12", marker = "(platform_machine == 'x86_64' and platform_system == 'Linux' and sys_platform == 'linux') or (platform_machine == 'x86_64' and platform_system == 'Linux' and sys_platform == 'windows')" }, - { name = "pytorch-triton", marker = "(python_full_version < '3.13' and platform_machine == 'x86_64' and platform_system == 'Linux' and sys_platform == 'linux') or (python_full_version < '3.13' and platform_machine == 'x86_64' and platform_system == 'Linux' and sys_platform == 'windows')" }, + { name = "pytorch-triton", marker = "(platform_machine == 'x86_64' and platform_system == 'Linux' and sys_platform == 'linux') or (platform_machine == 'x86_64' and platform_system == 'Linux' and sys_platform == 'windows')" }, { name = "setuptools", marker = "(python_full_version >= '3.12' and sys_platform == 'linux') or (python_full_version >= '3.12' and sys_platform == 'windows')" }, { name = "sympy", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, { name = "typing-extensions", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, ] wheels = [ - { url = "https://download.pytorch.org/whl/nightly/cu126/torch-2.6.0.dev20241212%2Bcu126-cp310-cp310-linux_aarch64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/cu126/torch-2.6.0.dev20241212%2Bcu126-cp310-cp310-manylinux_2_28_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/cu126/torch-2.6.0.dev20241212%2Bcu126-cp311-cp311-linux_aarch64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/cu126/torch-2.6.0.dev20241212%2Bcu126-cp311-cp311-manylinux_2_28_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/cu126/torch-2.6.0.dev20241212%2Bcu126-cp312-cp312-linux_aarch64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/cu126/torch-2.6.0.dev20241212%2Bcu126-cp312-cp312-manylinux_2_28_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/cu126/torch-2.6.0.dev20241212%2Bcu126-cp313-cp313-linux_aarch64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/cu126/torch-2.6.0.dev20241212%2Bcu126-cp313-cp313-manylinux_2_28_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/cu126/torch-2.6.0.dev20241212%2Bcu126-cp313-cp313t-manylinux_2_28_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/cu126/torch-2.6.0.dev20241212%2Bcu126-cp39-cp39-linux_aarch64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/cu126/torch-2.6.0.dev20241212%2Bcu126-cp39-cp39-manylinux_2_28_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/cu126/torch-2.6.0.dev20241224%2Bcu126-cp310-cp310-linux_aarch64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/cu126/torch-2.6.0.dev20241224%2Bcu126-cp310-cp310-manylinux_2_28_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/cu126/torch-2.6.0.dev20241224%2Bcu126-cp311-cp311-linux_aarch64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/cu126/torch-2.6.0.dev20241224%2Bcu126-cp311-cp311-manylinux_2_28_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/cu126/torch-2.6.0.dev20241224%2Bcu126-cp312-cp312-linux_aarch64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/cu126/torch-2.6.0.dev20241224%2Bcu126-cp312-cp312-manylinux_2_28_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/cu126/torch-2.6.0.dev20241224%2Bcu126-cp313-cp313-linux_aarch64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/cu126/torch-2.6.0.dev20241224%2Bcu126-cp313-cp313-manylinux_2_28_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/cu126/torch-2.6.0.dev20241224%2Bcu126-cp313-cp313t-manylinux_2_28_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/cu126/torch-2.6.0.dev20241224%2Bcu126-cp39-cp39-linux_aarch64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/cu126/torch-2.6.0.dev20241224%2Bcu126-cp39-cp39-manylinux_2_28_x86_64.whl" }, ] [[package]] name = "torch-tensorrt" -version = "2.6.0.dev0+302d0b8dd" +version = "2.6.0.dev0+b77a971dc" source = { editable = "." } dependencies = [ { name = "numpy", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, @@ -1336,6 +2432,9 @@ dependencies = [ ] [package.optional-dependencies] +distributed = [ + { name = "tensorrt-llm", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, +] jupyter = [ { name = "rich", extra = ["jupyter"], marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, ] @@ -1368,13 +2467,14 @@ dev = [ [package.metadata] requires-dist = [ { name = "numpy" }, - { name = "nvidia-modelopt", extras = ["deploy", "hf", "torch"], marker = "extra == 'quantization'", specifier = "~=0.17.0" }, + { name = "nvidia-modelopt", extras = ["deploy", "hf", "torch"], marker = "extra == 'quantization'", specifier = ">=0.17.0" }, { name = "packaging", specifier = ">=23" }, { name = "rich", marker = "extra == 'monitoring-tools'", specifier = ">=13.7.1" }, { name = "rich", extras = ["jupyter"], marker = "extra == 'jupyter'", specifier = ">=13.7.1" }, { name = "tensorrt-cu12", specifier = ">=10.6.0,<10.8.0" }, { name = "tensorrt-cu12-bindings", specifier = ">=10.6.0,<10.8.0" }, { name = "tensorrt-cu12-libs", specifier = ">=10.6.0,<10.8.0" }, + { name = "tensorrt-llm", marker = "extra == 'distributed'", specifier = ">=0.16.0" }, { name = "torch", specifier = ">=2.6.0.dev0,<2.7.0", index = "https://download.pytorch.org/whl/nightly/cu126" }, { name = "torchvision", marker = "extra == 'torchvision'", index = "https://download.pytorch.org/whl/nightly/cu126" }, { name = "typing-extensions", specifier = ">=4.7.0" }, @@ -1412,7 +2512,7 @@ wheels = [ [[package]] name = "torchvision" -version = "0.22.0.dev20241212+cu126" +version = "0.22.0.dev20241224+cu126" source = { registry = "https://download.pytorch.org/whl/nightly/cu126" } dependencies = [ { name = "numpy", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, @@ -1420,11 +2520,11 @@ dependencies = [ { name = "torch", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, ] wheels = [ - { url = "https://download.pytorch.org/whl/nightly/cu126/torchvision-0.22.0.dev20241212%2Bcu126-cp310-cp310-linux_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/cu126/torchvision-0.22.0.dev20241212%2Bcu126-cp311-cp311-linux_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/cu126/torchvision-0.22.0.dev20241212%2Bcu126-cp312-cp312-linux_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/cu126/torchvision-0.22.0.dev20241212%2Bcu126-cp313-cp313-linux_x86_64.whl" }, - { url = "https://download.pytorch.org/whl/nightly/cu126/torchvision-0.22.0.dev20241212%2Bcu126-cp39-cp39-linux_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/cu126/torchvision-0.22.0.dev20241224%2Bcu126-cp310-cp310-linux_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/cu126/torchvision-0.22.0.dev20241224%2Bcu126-cp311-cp311-linux_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/cu126/torchvision-0.22.0.dev20241224%2Bcu126-cp312-cp312-linux_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/cu126/torchvision-0.22.0.dev20241224%2Bcu126-cp313-cp313-linux_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/cu126/torchvision-0.22.0.dev20241224%2Bcu126-cp39-cp39-linux_x86_64.whl" }, ] [[package]] @@ -1449,7 +2549,7 @@ wheels = [ [[package]] name = "transformers" -version = "4.47.0" +version = "4.45.1" source = { registry = "https://pypi.org/simple" } dependencies = [ { name = "filelock", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, @@ -1463,9 +2563,9 @@ dependencies = [ { name = "tokenizers", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, { name = "tqdm", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, ] -sdist = { url = "https://files.pythonhosted.org/packages/b1/5a/0ecfde3264bed0579c37f249e04e15f3c1451ba864d78bbe390177664cac/transformers-4.47.0.tar.gz", hash = "sha256:f8ead7a5a4f6937bb507e66508e5e002dc5930f7b6122a9259c37b099d0f3b19", size = 8693668 } +sdist = { url = "https://files.pythonhosted.org/packages/15/1f/4d4c718c178b27268e5b9fa6112ebe82e190b703a984a51b0f743168ed49/transformers-4.45.1.tar.gz", hash = "sha256:9cace11072172df05ca6a694fcd1f5064a55b63285e492bd88f0ad1cec270f02", size = 8478117 } wheels = [ - { url = "https://files.pythonhosted.org/packages/d0/a7/7eedcf6a359e1e1eff3bc204ad022485aa5d88c08e1e3e0e0aee8a2e2235/transformers-4.47.0-py3-none-any.whl", hash = "sha256:a8e1bafdaae69abdda3cad638fe392e37c86d2ce0ecfcae11d60abb8f949ff4d", size = 10133426 }, + { url = "https://files.pythonhosted.org/packages/17/f2/f01ea29c8eff8e749d96525a17c2d3ec02656cec9a80c20fb3e74dba4b04/transformers-4.45.1-py3-none-any.whl", hash = "sha256:21e3f47aa7256dbbfb5215937a3168a984c94432ce3a16b7908265807d62aee8", size = 9881041 }, ] [[package]] @@ -1478,15 +2578,23 @@ wheels = [ [[package]] name = "typos" -version = "1.28.3" +version = "1.28.4" source = { registry = "https://pypi.org/simple" } -sdist = { url = "https://files.pythonhosted.org/packages/84/c2/ad5bade0f612955067a39bcd3b1998e94124e0a67cb9c50862284e3eafbd/typos-1.28.3.tar.gz", hash = "sha256:2fd9f5bcb6b6b0e98e302ae3c37e07a800888c0f510ef61e86ed2a6d60f64ea4", size = 1123306 } +sdist = { url = "https://files.pythonhosted.org/packages/52/26/b3c29fbf2726b8278578841a5f34ba244819660000b75d283a1afd1630a1/typos-1.28.4.tar.gz", hash = "sha256:7afd8ad79ab8b84f7adb12350d5630abc5e061c8a76802ddbc29eea256689600", size = 1127000 } wheels = [ - { url = "https://files.pythonhosted.org/packages/38/96/a4fae14932706132a9d15968083dbf1112df7aaafb635c4e358521d1206f/typos-1.28.3-py3-none-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:623d1b790fac1919ec65de8cfe4a4f0b283043024f9b13fa981b9c24d2829a20", size = 4313666 }, - { url = "https://files.pythonhosted.org/packages/20/4c/7fcdd933ff8faeaf6965c0b8a1b098b6b73945126a81118c690da807829f/typos-1.28.3-py3-none-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:96bc5323d09a306069fa6f586fd3544ccafa01a40acdd4ce2a3fa0023ff4b752", size = 3381866 }, - { url = "https://files.pythonhosted.org/packages/a2/99/899883f8781b98ff980ab8857a7b767fe3086d6018837e6c2de4fd4b0e59/typos-1.28.3-py3-none-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9dccfb721b2b4acb2bca613cc3f52d182f5f47c293a548110b0e2d198b89e050", size = 4096957 }, - { url = "https://files.pythonhosted.org/packages/6e/ee/73c0272c7af33939c182e1831b1c624cd16476b926f33f64dd477b68e318/typos-1.28.3-py3-none-musllinux_1_2_aarch64.whl", hash = "sha256:85de569b505e4d5ae6f7678a3df339eacab02f574934b8d6730ca87c065d9655", size = 3352137 }, - { url = "https://files.pythonhosted.org/packages/d4/8f/381dcdc66ac13202fa4588734b212f932d302bc54607cd67426da03339db/typos-1.28.3-py3-none-musllinux_1_2_x86_64.whl", hash = "sha256:a7df828acf62bdf0938d9fe59a270f6f0dc866028dbd876987d16acb9be5c64c", size = 4155022 }, + { url = "https://files.pythonhosted.org/packages/62/6e/84fe32eacaf2cc07647536f85c3d8329f804056d2c38b2294770b25a6c07/typos-1.28.4-py3-none-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bacfb01a2ab60b146f1412252f327e58e32a430613a761d76dbcc6e275ecffe3", size = 4530518 }, + { url = "https://files.pythonhosted.org/packages/9c/a8/ed642eff75832f1088a49414919d3d9323f1ea0ad467e65c7e122b2215c9/typos-1.28.4-py3-none-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b3bc312a10df63211b4f8730d551bc086f71ec5fb7a0a587a50f16c3902edf76", size = 3614918 }, + { url = "https://files.pythonhosted.org/packages/1c/89/974ef4925ba63aa5678157724eb96626b649e8601840821527e550b32a7f/typos-1.28.4-py3-none-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f605b3bb8c928cc0a0d46b29335d400630d43da0a9977bc890987a6cc175420a", size = 4339147 }, + { url = "https://files.pythonhosted.org/packages/63/ac/bbb69d77ae50106685a488aec676a6b43e1147dcd80ef0133a65c931fc54/typos-1.28.4-py3-none-musllinux_1_2_aarch64.whl", hash = "sha256:d854c5f854304efb959d7fe56fef5720163738687a6db6232bbd951ee2190167", size = 3567775 }, + { url = "https://files.pythonhosted.org/packages/48/0c/b7d85b3665d98e6f0eed10d534fafb3753a0712f07e86353be8cc86d2836/typos-1.28.4-py3-none-musllinux_1_2_x86_64.whl", hash = "sha256:ce87ddde847c535354dbe036691473fc6029f2c895f47340167874185b14bb29", size = 4400492 }, +] + +[[package]] +name = "tzdata" +version = "2024.2" +source = { registry = "https://download.pytorch.org/whl/nightly/cu126" } +wheels = [ + { url = "https://download.pytorch.org/whl/nightly/tzdata-2024.2-py2.py3-none-any.whl" }, ] [[package]] @@ -1497,6 +2605,20 @@ wheels = [ { url = "https://download.pytorch.org/whl/nightly/urllib3-2.2.3-py3-none-any.whl" }, ] +[[package]] +name = "uvicorn" +version = "0.34.0" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "click", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "h11", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "typing-extensions", marker = "(python_full_version < '3.11' and sys_platform == 'linux') or (python_full_version < '3.11' and sys_platform == 'windows')" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/4b/4d/938bd85e5bf2edeec766267a5015ad969730bb91e31b44021dfe8b22df6c/uvicorn-0.34.0.tar.gz", hash = "sha256:404051050cd7e905de2c9a7e61790943440b3416f49cb409f965d9dcd0fa73e9", size = 76568 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/61/14/33a3a1352cfa71812a3a21e8c9bfb83f60b0011f5e36f2b1399d51928209/uvicorn-0.34.0-py3-none-any.whl", hash = "sha256:023dc038422502fa28a09c7a30bf2b6991512da7dcdb8fd35fe57cfc154126f4", size = 62315 }, +] + [[package]] name = "virtualenv" version = "20.28.0" @@ -1520,6 +2642,15 @@ wheels = [ { url = "https://files.pythonhosted.org/packages/fd/84/fd2ba7aafacbad3c4201d395674fc6348826569da3c0937e75505ead3528/wcwidth-0.2.13-py2.py3-none-any.whl", hash = "sha256:3da69048e4540d84af32131829ff948f1e022c1c6bdb8d6102117aac784f6859", size = 34166 }, ] +[[package]] +name = "wheel" +version = "0.45.1" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/8a/98/2d9906746cdc6a6ef809ae6338005b3f21bb568bea3165cfc6a243fdc25c/wheel-0.45.1.tar.gz", hash = "sha256:661e1abd9198507b1409a20c02106d9670b2576e916d58f520316666abca6729", size = 107545 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/0b/2c/87f3254fd8ffd29e4c02732eee68a83a1d3c346ae39bc6822dcbcb697f2b/wheel-0.45.1-py3-none-any.whl", hash = "sha256:708e7481cc80179af0e556bbf0cc00b8444c7321e2700b8d8580231d13017248", size = 72494 }, +] + [[package]] name = "widgetsnbextension" version = "4.0.13" @@ -1529,6 +2660,107 @@ wheels = [ { url = "https://files.pythonhosted.org/packages/21/02/88b65cc394961a60c43c70517066b6b679738caf78506a5da7b88ffcb643/widgetsnbextension-4.0.13-py3-none-any.whl", hash = "sha256:74b2692e8500525cc38c2b877236ba51d34541e6385eeed5aec15a70f88a6c71", size = 2335872 }, ] +[[package]] +name = "xxhash" +version = "3.5.0" +source = { registry = "https://download.pytorch.org/whl/nightly/cu126" } +wheels = [ + { url = "https://download.pytorch.org/whl/nightly/xxhash-3.5.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/xxhash-3.5.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl" }, + { url = "https://download.pytorch.org/whl/nightly/xxhash-3.5.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl" }, + { url = "https://download.pytorch.org/whl/nightly/xxhash-3.5.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/xxhash-3.5.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl" }, + { url = "https://download.pytorch.org/whl/nightly/xxhash-3.5.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/xxhash-3.5.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl" }, + { url = "https://download.pytorch.org/whl/nightly/xxhash-3.5.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl" }, + { url = "https://download.pytorch.org/whl/nightly/xxhash-3.5.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/xxhash-3.5.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl" }, + { url = "https://download.pytorch.org/whl/nightly/xxhash-3.5.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/xxhash-3.5.0-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl" }, + { url = "https://download.pytorch.org/whl/nightly/xxhash-3.5.0-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl" }, + { url = "https://download.pytorch.org/whl/nightly/xxhash-3.5.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/xxhash-3.5.0-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl" }, + { url = "https://download.pytorch.org/whl/nightly/xxhash-3.5.0-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/xxhash-3.5.0-cp313-cp313-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl" }, + { url = "https://download.pytorch.org/whl/nightly/xxhash-3.5.0-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl" }, + { url = "https://download.pytorch.org/whl/nightly/xxhash-3.5.0-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/xxhash-3.5.0-cp313-cp313-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl" }, + { url = "https://download.pytorch.org/whl/nightly/xxhash-3.5.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/xxhash-3.5.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl" }, + { url = "https://download.pytorch.org/whl/nightly/xxhash-3.5.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl" }, + { url = "https://download.pytorch.org/whl/nightly/xxhash-3.5.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl" }, + { url = "https://download.pytorch.org/whl/nightly/xxhash-3.5.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl" }, +] + +[[package]] +name = "yarl" +version = "1.18.3" +source = { registry = "https://pypi.org/simple" } +dependencies = [ + { name = "idna", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "multidict", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, + { name = "propcache", marker = "sys_platform == 'linux' or sys_platform == 'windows'" }, +] +sdist = { url = "https://files.pythonhosted.org/packages/b7/9d/4b94a8e6d2b51b599516a5cb88e5bc99b4d8d4583e468057eaa29d5f0918/yarl-1.18.3.tar.gz", hash = "sha256:ac1801c45cbf77b6c99242eeff4fffb5e4e73a800b5c4ad4fc0be5def634d2e1", size = 181062 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/76/f9/d616a5c2daae281171de10fba41e1c0e2d8207166fc3547252f7d469b4e1/yarl-1.18.3-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c654d5207c78e0bd6d749f6dae1dcbbfde3403ad3a4b11f3c5544d9906969dde", size = 315349 }, + { url = "https://files.pythonhosted.org/packages/bb/b4/3ea5e7b6f08f698b3769a06054783e434f6d59857181b5c4e145de83f59b/yarl-1.18.3-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:5094d9206c64181d0f6e76ebd8fb2f8fe274950a63890ee9e0ebfd58bf9d787b", size = 330494 }, + { url = "https://files.pythonhosted.org/packages/55/f1/e0fc810554877b1b67420568afff51b967baed5b53bcc983ab164eebf9c9/yarl-1.18.3-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:35098b24e0327fc4ebdc8ffe336cee0a87a700c24ffed13161af80124b7dc8e5", size = 326927 }, + { url = "https://files.pythonhosted.org/packages/a9/42/b1753949b327b36f210899f2dd0a0947c0c74e42a32de3f8eb5c7d93edca/yarl-1.18.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3236da9272872443f81fedc389bace88408f64f89f75d1bdb2256069a8730ccc", size = 319703 }, + { url = "https://files.pythonhosted.org/packages/f0/6d/e87c62dc9635daefb064b56f5c97df55a2e9cc947a2b3afd4fd2f3b841c7/yarl-1.18.3-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e2c08cc9b16f4f4bc522771d96734c7901e7ebef70c6c5c35dd0f10845270bcd", size = 310246 }, + { url = "https://files.pythonhosted.org/packages/e3/ef/e2e8d1785cdcbd986f7622d7f0098205f3644546da7919c24b95790ec65a/yarl-1.18.3-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:80316a8bd5109320d38eef8833ccf5f89608c9107d02d2a7f985f98ed6876990", size = 319730 }, + { url = "https://files.pythonhosted.org/packages/fc/15/8723e22345bc160dfde68c4b3ae8b236e868f9963c74015f1bc8a614101c/yarl-1.18.3-cp310-cp310-musllinux_1_2_armv7l.whl", hash = "sha256:c1e1cc06da1491e6734f0ea1e6294ce00792193c463350626571c287c9a704db", size = 321681 }, + { url = "https://files.pythonhosted.org/packages/86/09/bf764e974f1516efa0ae2801494a5951e959f1610dd41edbfc07e5e0f978/yarl-1.18.3-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:fea09ca13323376a2fdfb353a5fa2e59f90cd18d7ca4eaa1fd31f0a8b4f91e62", size = 324812 }, + { url = "https://files.pythonhosted.org/packages/f6/4c/20a0187e3b903c97d857cf0272d687c1b08b03438968ae8ffc50fe78b0d6/yarl-1.18.3-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:e3b9fd71836999aad54084906f8663dffcd2a7fb5cdafd6c37713b2e72be1760", size = 337011 }, + { url = "https://files.pythonhosted.org/packages/c9/71/6244599a6e1cc4c9f73254a627234e0dad3883ece40cc33dce6265977461/yarl-1.18.3-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:757e81cae69244257d125ff31663249b3013b5dc0a8520d73694aed497fb195b", size = 338132 }, + { url = "https://files.pythonhosted.org/packages/af/f5/e0c3efaf74566c4b4a41cb76d27097df424052a064216beccae8d303c90f/yarl-1.18.3-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:b1771de9944d875f1b98a745bc547e684b863abf8f8287da8466cf470ef52690", size = 331849 }, + { url = "https://files.pythonhosted.org/packages/ed/fe/88b690b30f3f59275fb674f5f93ddd4a3ae796c2b62e5bb9ece8a4914b83/yarl-1.18.3-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d980e0325b6eddc81331d3f4551e2a333999fb176fd153e075c6d1c2530aa8a8", size = 340649 }, + { url = "https://files.pythonhosted.org/packages/07/eb/3b65499b568e01f36e847cebdc8d7ccb51fff716dbda1ae83c3cbb8ca1c9/yarl-1.18.3-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:b643562c12680b01e17239be267bc306bbc6aac1f34f6444d1bded0c5ce438ca", size = 356623 }, + { url = "https://files.pythonhosted.org/packages/33/46/f559dc184280b745fc76ec6b1954de2c55595f0ec0a7614238b9ebf69618/yarl-1.18.3-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c017a3b6df3a1bd45b9fa49a0f54005e53fbcad16633870104b66fa1a30a29d8", size = 354007 }, + { url = "https://files.pythonhosted.org/packages/af/ba/1865d85212351ad160f19fb99808acf23aab9a0f8ff31c8c9f1b4d671fc9/yarl-1.18.3-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:75674776d96d7b851b6498f17824ba17849d790a44d282929c42dbb77d4f17ae", size = 344145 }, + { url = "https://files.pythonhosted.org/packages/94/cb/5c3e975d77755d7b3d5193e92056b19d83752ea2da7ab394e22260a7b824/yarl-1.18.3-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ccaa3a4b521b780a7e771cc336a2dba389a0861592bbce09a476190bb0c8b4b3", size = 336133 }, + { url = "https://files.pythonhosted.org/packages/19/89/b77d3fd249ab52a5c40859815765d35c91425b6bb82e7427ab2f78f5ff55/yarl-1.18.3-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:2d06d3005e668744e11ed80812e61efd77d70bb7f03e33c1598c301eea20efbb", size = 347967 }, + { url = "https://files.pythonhosted.org/packages/35/bd/f6b7630ba2cc06c319c3235634c582a6ab014d52311e7d7c22f9518189b5/yarl-1.18.3-cp311-cp311-musllinux_1_2_armv7l.whl", hash = "sha256:9d41beda9dc97ca9ab0b9888cb71f7539124bc05df02c0cff6e5acc5a19dcc6e", size = 346397 }, + { url = "https://files.pythonhosted.org/packages/18/1a/0b4e367d5a72d1f095318344848e93ea70da728118221f84f1bf6c1e39e7/yarl-1.18.3-cp311-cp311-musllinux_1_2_i686.whl", hash = "sha256:ba23302c0c61a9999784e73809427c9dbedd79f66a13d84ad1b1943802eaaf59", size = 350206 }, + { url = "https://files.pythonhosted.org/packages/b5/cf/320fff4367341fb77809a2d8d7fe75b5d323a8e1b35710aafe41fdbf327b/yarl-1.18.3-cp311-cp311-musllinux_1_2_ppc64le.whl", hash = "sha256:6748dbf9bfa5ba1afcc7556b71cda0d7ce5f24768043a02a58846e4a443d808d", size = 362089 }, + { url = "https://files.pythonhosted.org/packages/57/cf/aadba261d8b920253204085268bad5e8cdd86b50162fcb1b10c10834885a/yarl-1.18.3-cp311-cp311-musllinux_1_2_s390x.whl", hash = "sha256:0b0cad37311123211dc91eadcb322ef4d4a66008d3e1bdc404808992260e1a0e", size = 366267 }, + { url = "https://files.pythonhosted.org/packages/54/58/fb4cadd81acdee6dafe14abeb258f876e4dd410518099ae9a35c88d8097c/yarl-1.18.3-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:0fb2171a4486bb075316ee754c6d8382ea6eb8b399d4ec62fde2b591f879778a", size = 359141 }, + { url = "https://files.pythonhosted.org/packages/6b/32/927b2d67a412c31199e83fefdce6e645247b4fb164aa1ecb35a0f9eb2058/yarl-1.18.3-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:436c4fc0a4d66b2badc6c5fc5ef4e47bb10e4fd9bf0c79524ac719a01f3607c2", size = 332368 }, + { url = "https://files.pythonhosted.org/packages/19/e5/859fca07169d6eceeaa4fde1997c91d8abde4e9a7c018e371640c2da2b71/yarl-1.18.3-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e35ef8683211db69ffe129a25d5634319a677570ab6b2eba4afa860f54eeaf75", size = 342314 }, + { url = "https://files.pythonhosted.org/packages/08/75/76b63ccd91c9e03ab213ef27ae6add2e3400e77e5cdddf8ed2dbc36e3f21/yarl-1.18.3-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:84b2deecba4a3f1a398df819151eb72d29bfeb3b69abb145a00ddc8d30094512", size = 341987 }, + { url = "https://files.pythonhosted.org/packages/1a/e1/a097d5755d3ea8479a42856f51d97eeff7a3a7160593332d98f2709b3580/yarl-1.18.3-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:00e5a1fea0fd4f5bfa7440a47eff01d9822a65b4488f7cff83155a0f31a2ecba", size = 336914 }, + { url = "https://files.pythonhosted.org/packages/0b/42/e1b4d0e396b7987feceebe565286c27bc085bf07d61a59508cdaf2d45e63/yarl-1.18.3-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d0e883008013c0e4aef84dcfe2a0b172c4d23c2669412cf5b3371003941f72bb", size = 325765 }, + { url = "https://files.pythonhosted.org/packages/7e/18/03a5834ccc9177f97ca1bbb245b93c13e58e8225276f01eedc4cc98ab820/yarl-1.18.3-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:5a3f356548e34a70b0172d8890006c37be92995f62d95a07b4a42e90fba54272", size = 344444 }, + { url = "https://files.pythonhosted.org/packages/c8/03/a713633bdde0640b0472aa197b5b86e90fbc4c5bc05b727b714cd8a40e6d/yarl-1.18.3-cp312-cp312-musllinux_1_2_armv7l.whl", hash = "sha256:ccd17349166b1bee6e529b4add61727d3f55edb7babbe4069b5764c9587a8cc6", size = 340760 }, + { url = "https://files.pythonhosted.org/packages/eb/99/f6567e3f3bbad8fd101886ea0276c68ecb86a2b58be0f64077396cd4b95e/yarl-1.18.3-cp312-cp312-musllinux_1_2_i686.whl", hash = "sha256:b958ddd075ddba5b09bb0be8a6d9906d2ce933aee81100db289badbeb966f54e", size = 346484 }, + { url = "https://files.pythonhosted.org/packages/8e/a9/84717c896b2fc6cb15bd4eecd64e34a2f0a9fd6669e69170c73a8b46795a/yarl-1.18.3-cp312-cp312-musllinux_1_2_ppc64le.whl", hash = "sha256:c7d79f7d9aabd6011004e33b22bc13056a3e3fb54794d138af57f5ee9d9032cb", size = 359864 }, + { url = "https://files.pythonhosted.org/packages/1e/2e/d0f5f1bef7ee93ed17e739ec8dbcb47794af891f7d165fa6014517b48169/yarl-1.18.3-cp312-cp312-musllinux_1_2_s390x.whl", hash = "sha256:4891ed92157e5430874dad17b15eb1fda57627710756c27422200c52d8a4e393", size = 364537 }, + { url = "https://files.pythonhosted.org/packages/97/8a/568d07c5d4964da5b02621a517532adb8ec5ba181ad1687191fffeda0ab6/yarl-1.18.3-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:ce1af883b94304f493698b00d0f006d56aea98aeb49d75ec7d98cd4a777e9285", size = 357861 }, + { url = "https://files.pythonhosted.org/packages/56/4e/d2563d8323a7e9a414b5b25341b3942af5902a2263d36d20fb17c40411e2/yarl-1.18.3-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:88a19f62ff30117e706ebc9090b8ecc79aeb77d0b1f5ec10d2d27a12bc9f66d0", size = 333587 }, + { url = "https://files.pythonhosted.org/packages/25/c9/cfec0bc0cac8d054be223e9f2c7909d3e8442a856af9dbce7e3442a8ec8d/yarl-1.18.3-cp313-cp313-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e17c9361d46a4d5addf777c6dd5eab0715a7684c2f11b88c67ac37edfba6c482", size = 344386 }, + { url = "https://files.pythonhosted.org/packages/ab/5d/4c532190113b25f1364d25f4c319322e86232d69175b91f27e3ebc2caf9a/yarl-1.18.3-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:1a74a13a4c857a84a845505fd2d68e54826a2cd01935a96efb1e9d86c728e186", size = 345421 }, + { url = "https://files.pythonhosted.org/packages/23/d1/6cdd1632da013aa6ba18cee4d750d953104a5e7aac44e249d9410a972bf5/yarl-1.18.3-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:41f7ce59d6ee7741af71d82020346af364949314ed3d87553763a2df1829cc58", size = 339384 }, + { url = "https://files.pythonhosted.org/packages/9a/c4/6b3c39bec352e441bd30f432cda6ba51681ab19bb8abe023f0d19777aad1/yarl-1.18.3-cp313-cp313-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f52a265001d830bc425f82ca9eabda94a64a4d753b07d623a9f2863fde532b53", size = 326689 }, + { url = "https://files.pythonhosted.org/packages/23/30/07fb088f2eefdc0aa4fc1af4e3ca4eb1a3aadd1ce7d866d74c0f124e6a85/yarl-1.18.3-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:82123d0c954dc58db301f5021a01854a85bf1f3bb7d12ae0c01afc414a882ca2", size = 345453 }, + { url = "https://files.pythonhosted.org/packages/63/09/d54befb48f9cd8eec43797f624ec37783a0266855f4930a91e3d5c7717f8/yarl-1.18.3-cp313-cp313-musllinux_1_2_armv7l.whl", hash = "sha256:2ec9bbba33b2d00999af4631a3397d1fd78290c48e2a3e52d8dd72db3a067ac8", size = 341872 }, + { url = "https://files.pythonhosted.org/packages/91/26/fd0ef9bf29dd906a84b59f0cd1281e65b0c3e08c6aa94b57f7d11f593518/yarl-1.18.3-cp313-cp313-musllinux_1_2_i686.whl", hash = "sha256:fbd6748e8ab9b41171bb95c6142faf068f5ef1511935a0aa07025438dd9a9bc1", size = 347497 }, + { url = "https://files.pythonhosted.org/packages/d9/b5/14ac7a256d0511b2ac168d50d4b7d744aea1c1aa20c79f620d1059aab8b2/yarl-1.18.3-cp313-cp313-musllinux_1_2_ppc64le.whl", hash = "sha256:877d209b6aebeb5b16c42cbb377f5f94d9e556626b1bfff66d7b0d115be88d0a", size = 359981 }, + { url = "https://files.pythonhosted.org/packages/ca/b3/d493221ad5cbd18bc07e642894030437e405e1413c4236dd5db6e46bcec9/yarl-1.18.3-cp313-cp313-musllinux_1_2_s390x.whl", hash = "sha256:b464c4ab4bfcb41e3bfd3f1c26600d038376c2de3297760dfe064d2cb7ea8e10", size = 366229 }, + { url = "https://files.pythonhosted.org/packages/04/56/6a3e2a5d9152c56c346df9b8fb8edd2c8888b1e03f96324d457e5cf06d34/yarl-1.18.3-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:8d39d351e7faf01483cc7ff7c0213c412e38e5a340238826be7e0e4da450fdc8", size = 360383 }, + { url = "https://files.pythonhosted.org/packages/0f/4f/438c9fd668954779e48f08c0688ee25e0673380a21bb1e8ccc56de5b55d7/yarl-1.18.3-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:09c7907c8548bcd6ab860e5f513e727c53b4a714f459b084f6580b49fa1b9cee", size = 317327 }, + { url = "https://files.pythonhosted.org/packages/bd/79/a78066f06179b4ed4581186c136c12fcfb928c475cbeb23743e71a991935/yarl-1.18.3-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:b4f6450109834af88cb4cc5ecddfc5380ebb9c228695afc11915a0bf82116789", size = 336999 }, + { url = "https://files.pythonhosted.org/packages/55/02/527963cf65f34a06aed1e766ff9a3b3e7d0eaa1c90736b2948a62e528e1d/yarl-1.18.3-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:a9ca04806f3be0ac6d558fffc2fdf8fcef767e0489d2684a21912cc4ed0cd1b8", size = 331693 }, + { url = "https://files.pythonhosted.org/packages/a2/2a/167447ae39252ba624b98b8c13c0ba35994d40d9110e8a724c83dbbb5822/yarl-1.18.3-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:77a6e85b90a7641d2e07184df5557132a337f136250caafc9ccaa4a2a998ca2c", size = 321473 }, + { url = "https://files.pythonhosted.org/packages/55/03/07955fabb20082373be311c91fd78abe458bc7ff9069d34385e8bddad20e/yarl-1.18.3-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6333c5a377c8e2f5fae35e7b8f145c617b02c939d04110c76f29ee3676b5f9a5", size = 313571 }, + { url = "https://files.pythonhosted.org/packages/95/e2/67c8d3ec58a8cd8ddb1d63bd06eb7e7b91c9f148707a3eeb5a7ed87df0ef/yarl-1.18.3-cp39-cp39-musllinux_1_2_aarch64.whl", hash = "sha256:0b3c92fa08759dbf12b3a59579a4096ba9af8dd344d9a813fc7f5070d86bbab1", size = 325004 }, + { url = "https://files.pythonhosted.org/packages/06/43/51ceb3e427368fe6ccd9eccd162be227fd082523e02bad1fd3063daf68da/yarl-1.18.3-cp39-cp39-musllinux_1_2_armv7l.whl", hash = "sha256:4ac515b860c36becb81bb84b667466885096b5fc85596948548b667da3bf9f24", size = 322677 }, + { url = "https://files.pythonhosted.org/packages/e4/0e/7ef286bfb23267739a703f7b967a858e2128c10bea898de8fa027e962521/yarl-1.18.3-cp39-cp39-musllinux_1_2_i686.whl", hash = "sha256:045b8482ce9483ada4f3f23b3774f4e1bf4f23a2d5c912ed5170f68efb053318", size = 332806 }, + { url = "https://files.pythonhosted.org/packages/c8/94/2d1f060f4bfa47c8bd0bcb652bfe71fba881564bcac06ebb6d8ced9ac3bc/yarl-1.18.3-cp39-cp39-musllinux_1_2_ppc64le.whl", hash = "sha256:a4bb030cf46a434ec0225bddbebd4b89e6471814ca851abb8696170adb163985", size = 339919 }, + { url = "https://files.pythonhosted.org/packages/8e/8d/73b5f9a6ab69acddf1ca1d5e7bc92f50b69124512e6c26b36844531d7f23/yarl-1.18.3-cp39-cp39-musllinux_1_2_s390x.whl", hash = "sha256:54d6921f07555713b9300bee9c50fb46e57e2e639027089b1d795ecd9f7fa910", size = 340960 }, + { url = "https://files.pythonhosted.org/packages/41/13/ce6bc32be4476b60f4f8694831f49590884b2c975afcffc8d533bf2be7ec/yarl-1.18.3-cp39-cp39-musllinux_1_2_x86_64.whl", hash = "sha256:1d407181cfa6e70077df3377938c08012d18893f9f20e92f7d2f314a437c30b1", size = 336592 }, + { url = "https://files.pythonhosted.org/packages/f5/4b/a06e0ec3d155924f77835ed2d167ebd3b211a7b0853da1cf8d8414d784ef/yarl-1.18.3-py3-none-any.whl", hash = "sha256:b57f4f58099328dfb26c6a771d09fb20dbbae81d20cfb66141251ea063bd101b", size = 45109 }, +] + [[package]] name = "zipp" version = "3.19.2" diff --git a/version.txt b/version.txt index 3d87ca93f8..787e6e4ab7 100644 --- a/version.txt +++ b/version.txt @@ -1 +1 @@ -2.6.0a0 +2.7.0a0