Jetson Inference
DNN Vision Library
|
#include <NvInfer.h>
#include <jetson-utils/cudaUtility.h>
#include <jetson-utils/commandLine.h>
#include <jetson-utils/imageFormat.h>
#include <jetson-utils/timespec.h>
#include <jetson-utils/logging.h>
#include <vector>
#include <sstream>
#include <math.h>
Go to the source code of this file.
Classes | |
class | tensorNet |
Abstract class for loading a tensor network with TensorRT. More... | |
class | tensorNet::Logger |
Logger class for GIE info/warning/errors. More... | |
class | tensorNet::Profiler |
Profiler interface for measuring layer timings. More... | |
struct | tensorNet::layerInfo |
Namespaces | |
nvinfer1 | |
Macros | |
#define | DIMS_C(x) x.c |
#define | DIMS_H(x) x.h |
#define | DIMS_W(x) x.w |
#define | NV_TENSORRT_MAJOR 1 |
#define | NV_TENSORRT_MINOR 0 |
#define | NOEXCEPT |
#define | TENSORRT_VERSION_CHECK(major, minor, patch) (NV_TENSORRT_MAJOR > major || (NV_TENSORRT_MAJOR == major && NV_TENSORRT_MINOR > minor) || (NV_TENSORRT_MAJOR == major && NV_TENSORRT_MINOR == minor && NV_TENSORRT_PATCH >= patch)) |
Macro for checking the minimum version of TensorRT that is installed. More... | |
#define | DEFAULT_MAX_BATCH_SIZE 1 |
Default maximum batch size. More... | |
#define | LOG_TRT "[TRT] " |
Prefix used for tagging printed log output from TensorRT. More... | |
Typedefs | |
typedef nvinfer1::Dims3 | Dims3 |
Enumerations | |
enum | precisionType { TYPE_DISABLED = 0, TYPE_FASTEST, TYPE_FP32, TYPE_FP16, TYPE_INT8, NUM_PRECISIONS } |
Enumeration for indicating the desired precision that the network should run in, if available in hardware. More... | |
enum | deviceType { DEVICE_GPU = 0, DEVICE_DLA, DEVICE_DLA_0 = DEVICE_DLA, DEVICE_DLA_1, NUM_DEVICES } |
Enumeration for indicating the desired device that the network should run on, if available in hardware. More... | |
enum | modelType { MODEL_CUSTOM = 0, MODEL_CAFFE, MODEL_ONNX, MODEL_UFF, MODEL_ENGINE } |
Enumeration indicating the format of the model that's imported in TensorRT (either caffe, ONNX, or UFF). More... | |
enum | profilerQuery { PROFILER_PREPROCESS = 0, PROFILER_NETWORK, PROFILER_POSTPROCESS, PROFILER_VISUALIZE, PROFILER_TOTAL } |
Profiling queries. More... | |
enum | profilerDevice { PROFILER_CPU = 0, PROFILER_CUDA } |
Profiler device. More... | |
Functions | |
const char * | precisionTypeToStr (precisionType type) |
Stringize function that returns precisionType in text. More... | |
precisionType | precisionTypeFromStr (const char *str) |
Parse the precision type from a string. More... | |
const char * | deviceTypeToStr (deviceType type) |
Stringize function that returns deviceType in text. More... | |
deviceType | deviceTypeFromStr (const char *str) |
Parse the device type from a string. More... | |
const char * | modelTypeToStr (modelType type) |
Stringize function that returns modelType in text. More... | |
modelType | modelTypeFromStr (const char *str) |
Parse the model format from a string. More... | |
modelType | modelTypeFromPath (const char *path) |
Parse the model format from a file path. More... | |
const char * | profilerQueryToStr (profilerQuery query) |
Stringize function that returns profilerQuery in text. More... | |
#define DIMS_C | ( | x | ) | x.c |
#define DIMS_H | ( | x | ) | x.h |
#define DIMS_W | ( | x | ) | x.w |
#define NOEXCEPT |
#define NV_TENSORRT_MAJOR 1 |
#define NV_TENSORRT_MINOR 0 |
typedef nvinfer1::Dims3 Dims3 |