onnx.hub#
ModelInfo#
- class onnx.hub.ModelInfo(raw_model_info: dict[str, Any])[source]#
A class to represent a model’s property and metadata in the ONNX Hub. It extracts model name, path, sha, tags, etc. from the passed in raw_model_info dict.
- model#
The name of the model.
- model_path#
The path to the model, relative to the model zoo (https://github.com/onnx/models/) repo root.
- metadata#
Additional metadata of the model, such as the size of the model, IO ports, etc.
- model_sha#
The SHA256 digest of the model file.
- tags#
A set of tags associated with the model.
- opset#
The opset version of the model.
download_model_with_test_data#
- onnx.hub.download_model_with_test_data(model: str, repo: str = 'onnx/models:main', opset: int | None = None, force_reload: bool = False, silent: bool = False) str | None [source]#
Downloads a model along with test data by name from the onnx model hub and returns the directory to which the files have been extracted.
- Parameters:
model – The name of the onnx model in the manifest. This field is case-sensitive
repo – The location of the model repo in format “user/repo[:branch]”. If no branch is found will default to “main”
opset – The opset of the model to download. The default of None automatically chooses the largest opset
force_reload – Whether to force the model to re-download even if its already found in the cache
silent – Whether to suppress the warning message if the repo is not trusted.
- Returns:
str or None
get_model_info#
- onnx.hub.get_model_info(model: str, repo: str = 'onnx/models:main', opset: int | None = None) ModelInfo [source]#
Gets the model info matching the given name and opset.
- Parameters:
model – The name of the onnx model in the manifest. This field is case-sensitive
repo – The location of the model repo in format “user/repo[:branch]”. If no branch is found will default to “main”
opset – The opset of the model to get. The default of None will return the model with largest opset.
- Returns:
ModelInfo
.
list_models#
- onnx.hub.list_models(repo: str = 'onnx/models:main', model: str | None = None, tags: list[str] | None = None) list[ModelInfo] [source]#
Gets the list of model info consistent with a given name and tags
- Parameters:
repo – The location of the model repo in format “user/repo[:branch]”. If no branch is found will default to “main”
model – The name of the model to search for. If None, will return all models with matching tags.
tags – A list of tags to filter models by. If None, will return all models with matching name.
- Returns:
``ModelInfo``s.
load#
- onnx.hub.load(model: str, repo: str = 'onnx/models:main', opset: int | None = None, force_reload: bool = False, silent: bool = False) ModelProto | None [source]#
Downloads a model by name from the onnx model hub.
- Parameters:
model – The name of the onnx model in the manifest. This field is case-sensitive
repo – The location of the model repo in format “user/repo[:branch]”. If no branch is found will default to “main”
opset – The opset of the model to download. The default of None automatically chooses the largest opset
force_reload – Whether to force the model to re-download even if its already found in the cache
silent – Whether to suppress the warning message if the repo is not trusted.
- Returns:
ModelProto or None
load_composite_model#
- onnx.hub.load_composite_model(network_model: str, preprocessing_model: str, network_repo: str = 'onnx/models:main', preprocessing_repo: str = 'onnx/models:main', opset: int | None = None, force_reload: bool = False, silent: bool = False) ModelProto | None [source]#
Builds a composite model including data preprocessing by downloading a network and a preprocessing model and combine it into a single model
- Parameters:
network_model – The name of the onnx model in the manifest.
preprocessing_model – The name of the preprocessing model.
network_repo – The location of the model repo in format “user/repo[:branch]”. If no branch is found will default to “main”
preprocessing_repo – The location of the proprocessing model repo in format “user/repo[:branch]”. If no branch is found will default to “main”
opset – The opset of the model to download. The default of None automatically chooses the largest opset
force_reload – Whether to force the model to re-download even if its already found in the cache
silent – Whether to suppress the warning message if the repo is not trusted.
- Returns:
ModelProto or None