![]()  | 
  
    Jetson Inference
    
   DNN Vision Library 
   | 
 
#include <NvInfer.h>#include <jetson-utils/cudaUtility.h>#include <jetson-utils/commandLine.h>#include <jetson-utils/imageFormat.h>#include <jetson-utils/timespec.h>#include <jetson-utils/logging.h>#include <vector>#include <sstream>#include <math.h>Go to the source code of this file.
Classes | |
| class | tensorNet | 
| Abstract class for loading a tensor network with TensorRT.  More... | |
| class | tensorNet::Logger | 
| Logger class for GIE info/warning/errors.  More... | |
| class | tensorNet::Profiler | 
| Profiler interface for measuring layer timings.  More... | |
| struct | tensorNet::layerInfo | 
Namespaces | |
| nvinfer1 | |
Macros | |
| #define | DIMS_C(x) x.c | 
| #define | DIMS_H(x) x.h | 
| #define | DIMS_W(x) x.w | 
| #define | NV_TENSORRT_MAJOR 1 | 
| #define | NV_TENSORRT_MINOR 0 | 
| #define | NOEXCEPT | 
| #define | TENSORRT_VERSION_CHECK(major, minor, patch) (NV_TENSORRT_MAJOR > major || (NV_TENSORRT_MAJOR == major && NV_TENSORRT_MINOR > minor) || (NV_TENSORRT_MAJOR == major && NV_TENSORRT_MINOR == minor && NV_TENSORRT_PATCH >= patch)) | 
| Macro for checking the minimum version of TensorRT that is installed.  More... | |
| #define | DEFAULT_MAX_BATCH_SIZE 1 | 
| Default maximum batch size.  More... | |
| #define | LOG_TRT "[TRT] " | 
| Prefix used for tagging printed log output from TensorRT.  More... | |
Typedefs | |
| typedef nvinfer1::Dims3 | Dims3 | 
Enumerations | |
| enum | precisionType {  TYPE_DISABLED = 0, TYPE_FASTEST, TYPE_FP32, TYPE_FP16, TYPE_INT8, NUM_PRECISIONS }  | 
| Enumeration for indicating the desired precision that the network should run in, if available in hardware.  More... | |
| enum | deviceType {  DEVICE_GPU = 0, DEVICE_DLA, DEVICE_DLA_0 = DEVICE_DLA, DEVICE_DLA_1, NUM_DEVICES }  | 
| Enumeration for indicating the desired device that the network should run on, if available in hardware.  More... | |
| enum | modelType {  MODEL_CUSTOM = 0, MODEL_CAFFE, MODEL_ONNX, MODEL_UFF, MODEL_ENGINE }  | 
| Enumeration indicating the format of the model that's imported in TensorRT (either caffe, ONNX, or UFF).  More... | |
| enum | profilerQuery {  PROFILER_PREPROCESS = 0, PROFILER_NETWORK, PROFILER_POSTPROCESS, PROFILER_VISUALIZE, PROFILER_TOTAL }  | 
| Profiling queries.  More... | |
| enum | profilerDevice { PROFILER_CPU = 0, PROFILER_CUDA } | 
| Profiler device.  More... | |
Functions | |
| const char * | precisionTypeToStr (precisionType type) | 
| Stringize function that returns precisionType in text.  More... | |
| precisionType | precisionTypeFromStr (const char *str) | 
| Parse the precision type from a string.  More... | |
| const char * | deviceTypeToStr (deviceType type) | 
| Stringize function that returns deviceType in text.  More... | |
| deviceType | deviceTypeFromStr (const char *str) | 
| Parse the device type from a string.  More... | |
| const char * | modelTypeToStr (modelType type) | 
| Stringize function that returns modelType in text.  More... | |
| modelType | modelTypeFromStr (const char *str) | 
| Parse the model format from a string.  More... | |
| modelType | modelTypeFromPath (const char *path) | 
| Parse the model format from a file path.  More... | |
| const char * | profilerQueryToStr (profilerQuery query) | 
| Stringize function that returns profilerQuery in text.  More... | |
| #define DIMS_C | ( | x | ) | x.c | 
| #define DIMS_H | ( | x | ) | x.h | 
| #define DIMS_W | ( | x | ) | x.w | 
| #define NOEXCEPT | 
| #define NV_TENSORRT_MAJOR 1 | 
| #define NV_TENSORRT_MINOR 0 | 
| typedef nvinfer1::Dims3 Dims3 |