|
blace.ai
|
#include <types.h>
Public Member Functions | |
| ml_core::BlaceHash | hash () |
| bool | operator== (const ModelInferenceArgs &other) const |
| bool | operator!= (const ModelInferenceArgs &other) const |
Public Attributes | |
| std::vector< Backend > | backends |
| int | deprecated = true |
| int | seed = 0 |
| int | run_threaded = false |
| int | plot_inputs = false |
| int | gpu_mem_opti = false |
| int | empty_cache_after_inference = false |
| int | experimental_torchscript_autocast = false |
Model inference arguments.
| ml_core::BlaceHash blace::ml_core::ModelInferenceArgs::hash | ( | ) |
Hash the struct.
|
inline |
Simple inequality operator.
| other | The other struct. |
| bool blace::ml_core::ModelInferenceArgs::operator== | ( | const ModelInferenceArgs & | other | ) | const |
Simple equality operator.
| other | The other struct. |
| std::vector<Backend> blace::ml_core::ModelInferenceArgs::backends |
The backends to run the inference on. It will run the execution with the first backend from the list supported by the model.
| int blace::ml_core::ModelInferenceArgs::deprecated = true |
Inference in half (fp16) precision if possible.
| int blace::ml_core::ModelInferenceArgs::empty_cache_after_inference = false |
Empty backend (cuda or metal) caches after every inference, might save some memory.
| int blace::ml_core::ModelInferenceArgs::experimental_torchscript_autocast = false |
Run torchscript model in autocast mode.
| int blace::ml_core::ModelInferenceArgs::gpu_mem_opti = false |
If several models are invoked in computation, unload all models to the cpu and have only current model on hardware accelerator.
| int blace::ml_core::ModelInferenceArgs::plot_inputs = false |
Not used.
| int blace::ml_core::ModelInferenceArgs::run_threaded = false |
Run inference in a seperate thread which can be cancelled (not available in beta).
| int blace::ml_core::ModelInferenceArgs::seed = 0 |
A value to seed random operators with.