![]() |
Jetson Inference
DNN Vision Library
|
This is the complete list of members for homographyNet, including all inherited members.
| AllowGPUFallback() const | tensorNet | inline |
| COCO_128 enum value | homographyNet | |
| ComputeHomography(const float displacement[8], float H[3][3]) | homographyNet | |
| ComputeHomography(const float displacement[8], float H[3][3], float H_inv[3][3]) | homographyNet | |
| Create(NetworkType networkType=WEBCAM_320, uint32_t maxBatchSize=1, precisionType precision=TYPE_FASTEST, deviceType device=DEVICE_GPU, bool allowGPUFallback=true) | homographyNet | static |
| Create(const char *model_path, const char *input=HOMOGRAPHY_NET_DEFAULT_INPUT, const char *output=HOMOGRAPHY_NET_DEFAULT_OUTPUT, uint32_t maxBatchSize=1, precisionType precision=TYPE_FASTEST, deviceType device=DEVICE_GPU, bool allowGPUFallback=true) | homographyNet | static |
| Create(int argc, char **argv) | homographyNet | static |
| CreateStream(bool nonBlocking=true) | tensorNet | |
| CUSTOM enum value | homographyNet | |
| DetectNativePrecision(const std::vector< precisionType > &nativeTypes, precisionType type) | tensorNet | static |
| DetectNativePrecision(precisionType precision, deviceType device=DEVICE_GPU) | tensorNet | static |
| DetectNativePrecisions(deviceType device=DEVICE_GPU) | tensorNet | static |
| EnableDebug() | tensorNet | |
| EnableLayerProfiler() | tensorNet | |
| FindDisplacement(float *imageA, float *imageB, uint32_t width, uint32_t height, float displacement[8]) | homographyNet | |
| FindFastestPrecision(deviceType device=DEVICE_GPU, bool allowInt8=true) | tensorNet | static |
| FindHomography(float *imageA, float *imageB, uint32_t width, uint32_t height, float H[3][3]) | homographyNet | |
| FindHomography(float *imageA, float *imageB, uint32_t width, uint32_t height, float H[3][3], float H_inv[3][3]) | homographyNet | |
| GetDevice() const | tensorNet | inline |
| GetModelPath() const | tensorNet | inline |
| GetModelType() const | tensorNet | inline |
| GetNetworkTime() | tensorNet | inline |
| GetPrecision() const | tensorNet | inline |
| GetProfilerTime(profilerQuery query) | tensorNet | inline |
| GetProfilerTime(profilerQuery query, profilerDevice device) | tensorNet | inline |
| GetPrototxtPath() const | tensorNet | inline |
| GetStream() const | tensorNet | inline |
| gLogger | tensorNet | protected |
| gProfiler | tensorNet | protected |
| homographyNet() | homographyNet | protected |
| IsModelType(modelType type) const | tensorNet | inline |
| IsPrecision(precisionType type) const | tensorNet | inline |
| LoadNetwork(const char *prototxt, const char *model, const char *mean=NULL, const char *input_blob="data", const char *output_blob="prob", uint32_t maxBatchSize=DEFAULT_MAX_BATCH_SIZE, precisionType precision=TYPE_FASTEST, deviceType device=DEVICE_GPU, bool allowGPUFallback=true, nvinfer1::IInt8Calibrator *calibrator=NULL, cudaStream_t stream=NULL) | tensorNet | |
| LoadNetwork(const char *prototxt, const char *model, const char *mean, const char *input_blob, const std::vector< std::string > &output_blobs, uint32_t maxBatchSize=DEFAULT_MAX_BATCH_SIZE, precisionType precision=TYPE_FASTEST, deviceType device=DEVICE_GPU, bool allowGPUFallback=true, nvinfer1::IInt8Calibrator *calibrator=NULL, cudaStream_t stream=NULL) | tensorNet | |
| LoadNetwork(const char *prototxt, const char *model, const char *mean, const char *input_blob, const Dims3 &input_dims, const std::vector< std::string > &output_blobs, uint32_t maxBatchSize=DEFAULT_MAX_BATCH_SIZE, precisionType precision=TYPE_FASTEST, deviceType device=DEVICE_GPU, bool allowGPUFallback=true, nvinfer1::IInt8Calibrator *calibrator=NULL, cudaStream_t stream=NULL) | tensorNet | |
| mAllowGPUFallback | tensorNet | protected |
| mCacheCalibrationPath | tensorNet | protected |
| mCacheEnginePath | tensorNet | protected |
| mContext | tensorNet | protected |
| mDevice | tensorNet | protected |
| mEnableDebug | tensorNet | protected |
| mEnableProfiler | tensorNet | protected |
| mEngine | tensorNet | protected |
| mEventsCPU | tensorNet | protected |
| mEventsGPU | tensorNet | protected |
| mHeight | tensorNet | protected |
| mInfer | tensorNet | protected |
| mInputBlobName | tensorNet | protected |
| mInputCPU | tensorNet | protected |
| mInputCUDA | tensorNet | protected |
| mInputDims | tensorNet | protected |
| mInputSize | tensorNet | protected |
| mMaxBatchSize | tensorNet | protected |
| mMeanPath | tensorNet | protected |
| mModelPath | tensorNet | protected |
| mModelType | tensorNet | protected |
| mOutputs | tensorNet | protected |
| mPrecision | tensorNet | protected |
| mProfilerQueriesDone | tensorNet | protected |
| mProfilerQueriesUsed | tensorNet | protected |
| mProfilerTimes | tensorNet | protected |
| mPrototxtPath | tensorNet | protected |
| mStream | tensorNet | protected |
| mWidth | tensorNet | protected |
| NetworkType enum name | homographyNet | |
| NetworkTypeFromStr(const char *model_name) | homographyNet | static |
| PrintProfilerTimes() | tensorNet | inline |
| ProfileModel(const std::string &deployFile, const std::string &modelFile, const char *input, const Dims3 &inputDims, const std::vector< std::string > &outputs, uint32_t maxBatchSize, precisionType precision, deviceType device, bool allowGPUFallback, nvinfer1::IInt8Calibrator *calibrator, std::ostream &modelStream) | tensorNet | protected |
| PROFILER_BEGIN(profilerQuery query) | tensorNet | inlineprotected |
| PROFILER_END(profilerQuery query) | tensorNet | inlineprotected |
| PROFILER_QUERY(profilerQuery query) | tensorNet | inlineprotected |
| SetStream(cudaStream_t stream) | tensorNet | |
| tensorNet() | tensorNet | protected |
| WEBCAM_320 enum value | homographyNet | |
| ~homographyNet() | homographyNet | virtual |
| ~tensorNet() | tensorNet | virtual |