Profile Options
- class TfliteDelegates(value)[source]
Bases:
str
,Enum
An enumeration.
- QNN = 'qnn'
- QNN_GPU = 'qnn-gpu'
- NNAPI = 'nnapi'
- NNAPI_GPU = 'nnapi-gpu'
- GPU = 'gpu'
- XNNPACK = 'xnnpack'
- class ExecutionMode(value)[source]
Bases:
str
,Enum
An enumeration.
- SEQUENTIAL = 'SEQUENTIAL'
- PARALLEL = 'PARALLEL'
- class GraphOptimizationLevel(value)[source]
Bases:
str
,Enum
An enumeration.
- DISABLE_ALL = 'DISABLE_ALL'
- ENABLE_BASIC = 'ENABLE_BASIC'
- ENABLE_EXTENDED = 'ENABLE_EXTENDED'
- ENABLE_ALL = 'ENABLE_ALL'
- class OnnxQnnHtpPerformanceMode(value)[source]
Bases:
str
,Enum
An enumeration.
- DEFAULT = 'default'
- LOW_POWER_SAVER = 'low_power_saver'
- POWER_SAVER = 'power_saver'
- HIGH_POWER_SAVER = 'high_power_saver'
- LOW_BALANCED = 'low_balanced'
- BALANCED = 'balanced'
- HIGH_PERFORMANCE = 'high_performance'
- SUSTAINED_HIGH_PERFORMANCE = 'sustained_high_performance'
- BURST = 'burst'
- class OnnxExecutionProviders(value)[source]
Bases:
str
,Enum
An enumeration.
- QNN = 'qnn'
- QNN_GPU = 'qnn-gpu'
- DIRECTML = 'directml'
- class QnnLogLevel(value)[source]
Bases:
str
,Enum
An enumeration.
- K_LOG_OFF = 'kLogOff'
- K_LOG_LEVEL_ERROR = 'kLogLevelError'
- K_LOG_LEVEL_WARN = 'kLogLevelWarn'
- K_LOG_LEVEL_INFO = 'kLogLevelInfo'
- K_LOG_LEVEL_VERBOSE = 'kLogLevelVerbose'
- K_LOG_LEVEL_DEBUG = 'kLogLevelDebug'
- class QnnGraphPriority(value)[source]
Bases:
str
,Enum
An enumeration.
- K_QNN_PRIORITY_DEFAULT = 'kQnnPriorityDefault'
- K_QNN_PRIORITY_LOW = 'kQnnPriorityLow'
- K_QNN_PRIORITY_NORMAL = 'kQnnPriorityNormal'
- K_QNN_PRIORITY_NORMAL_HIGH = 'kQnnPriorityNormalHigh'
- K_QNN_PRIORITY_HIGH = 'kQnnPriorityHigh'
- K_QNN_PRIORITY_UNDEFINED = 'kQnnPriorityUndefined'
- class QnnGpuPrecision(value)[source]
Bases:
str
,Enum
An enumeration.
- K_GPU_USER_PROVIDED = 'kGpuUserProvided'
- K_GPU_FP32 = 'kGpuFp32'
- K_GPU_FP16 = 'kGpuFp16'
- K_GPU_HYBRID = 'kGpuHybrid'
- class QnnGpuPerformanceMode(value)[source]
Bases:
str
,Enum
An enumeration.
- K_GPU_DEFAULT = 'kGpuDefault'
- K_GPU_HIGH = 'kGpuHigh'
- K_GPU_NORMAL = 'kGpuNormal'
- K_GPU_LOW = 'kGpuLow'
- class QnnDspPerformanceMode(value)[source]
Bases:
str
,Enum
An enumeration.
- K_DSP_LOW_POWER_SAVER = 'kDspLowPowerSaver'
- K_DSP_POWER_SAVER = 'kDspPowerSaver'
- K_DSP_HIGH_POWER_SAVER = 'kDspHighPowerSaver'
- K_DSP_LOW_BALANCED = 'kDspLowBalanced'
- K_DSP_BALANCED = 'kDspBalanced'
- K_DSP_HIGH_PERFORMANCE = 'kDspHighPerformance'
- K_DPS_SUSTAINED_HIGH_PERFORMANCE = 'kDspSustainedHighPerformance'
- K_DSP_BURST = 'kDspBurst'
- class QnnDspEncoding(value)[source]
Bases:
str
,Enum
An enumeration.
- K_DSP_STATIC = 'kDspStatic'
- K_DSP_DYNAMIC = 'kDspDynamic'
- class TfliteQnnHtpPerformanceMode(value)[source]
Bases:
str
,Enum
An enumeration.
- K_HTP_LOW_POWER_SAVER = 'kHtpLowPowerSaver'
- K_HTP_POWER_SAVER = 'kHtpPowerSaver'
- K_HTP_HIGH_POWER_SAVER = 'kHtpHighPowerSaver'
- K_HTP_LOW_BALANCED = 'kHtpLowBalanced'
- K_HTP_BALANCED = 'kHtpBalanced'
- K_HTP_HIGH_PERFORMANCE = 'kHtpHighPerformance'
- K_HTP_SUSTAINED_HIGH_PERFORMANCE = 'kHtpSustainedHighPerformance'
- K_HTP_BURST = 'kHtpBurst'
- class QnnHtpPrecision(value)[source]
Bases:
str
,Enum
An enumeration.
- K_HTP_QUANTIZED = 'kHtpQuantized'
- K_HTP_FP16 = 'kHtpFp16'
- class QnnHtpOptimizationStrategy(value)[source]
Bases:
str
,Enum
An enumeration.
- K_HTP_OPTIMIZE_FOR_INFERENCE = 'kHtpOptimizeForInference'
- K_HTP_OPTIMIZE_FOR_PREPARE = 'kHtpOptimizeForPrepare'
- class GpuInferencePreference(value)[source]
Bases:
str
,Enum
An enumeration.
- TFLITE_GPU_INFERENCE_PREFERENCE_FAST_SINGLE_ANSWER = 'TFLITE_GPU_INFERENCE_PREFERENCE_FAST_SINGLE_ANSWER'
- TFLITE_GPU_INFERENCE_PREFERENCE_SUSTAINED_SPEED = 'TFLITE_GPU_INFERENCE_PREFERENCE_SUSTAINED_SPEED'
- TFLITE_GPU_INFERENCE_PREFERENCE_BALANCED = 'TFLITE_GPU_INFERENCE_PREFERENCE_BALANCED'
- class GpuInferencePriority(value)[source]
Bases:
str
,Enum
An enumeration.
- TFLITE_GPU_INFERENCE_PREFERENCE_BALANCED = 'TFLITE_GPU_INFERENCE_PREFERENCE_BALANCED'
- TFLITE_GPU_INFERENCE_PRIORITY_MAX_PRECISION = 'TFLITE_GPU_INFERENCE_PRIORITY_MAX_PRECISION'
- TFLITE_GPU_INFERENCE_PRIORITY_MIN_LATENCY = 'TFLITE_GPU_INFERENCE_PRIORITY_MIN_LATENCY'
- TFLITE_GPU_INFERENCE_PRIORITY_MIN_MEMORY_USAGE = 'TFLITE_GPU_INFERENCE_PRIORITY_MIN_MEMORY_USAGE'
- class NnapiExecutionPreference(value)[source]
Bases:
str
,Enum
An enumeration.
- K_LOW_POWER = 'kLowPower'
- K_FAST_SINGLE_ANSWER = 'kFastSingleAnswer'
- K_SUSTAINED_SPEED = 'kSustainedSpeed'
- class ContextErrorReportingOptionsLevel(value)[source]
Bases:
str
,Enum
An enumeration.
- BRIEF = 'BRIEF'
- DETAILED = 'DETAILED'
- class Priority(value)[source]
Bases:
str
,Enum
An enumeration.
- LOW = 'LOW'
- NORMAL = 'NORMAL'
- NORMAL_HIGH = 'NORMAL_HIGH'
- HIGH = 'HIGH'
- class ContextGpuPerformanceHint(value)[source]
Bases:
str
,Enum
An enumeration.
- LOW = 'LOW'
- NORMAL = 'NORMAL'
- HIGH = 'HIGH'
- class ContextHtpPerformanceMode(value)[source]
Bases:
str
,Enum
An enumeration.
- EXTREME_POWER_SAVER = 'EXTREME_POWER_SAVER'
- LOW_POWER_SAVER = 'LOW_POWER_SAVER'
- POWER_SAVER = 'POWER_SAVER'
- HIGH_POWER_SAVER = 'HIGH_POWER_SAVER'
- LOW_BALANCED = 'LOW_BALANCED'
- BALANCED = 'BALANCED'
- HIGH_PERFORMANCE = 'HIGH_PERFORMANCE'
- SUSTAINED_HIGH_PERFORMANCE = 'SUSTAINED_HIGH_PERFORMANCE'
- BURST = 'BURST'
- class DefaultGraphGpuPrecision(value)[source]
Bases:
str
,Enum
An enumeration.
- FLOAT32 = 'FLOAT32'
- FLOAT16 = 'FLOAT16'
- HYBRID = 'HYBRID'
- USER_PROVIDED = 'USER_PROVIDED'
- class DefaultGraphHtpOptimizationType(value)[source]
Bases:
str
,Enum
An enumeration.
- FINALIZE_OPTIMIZATION_FLAG = 'FINALIZE_OPTIMIZATION_FLAG'
- class DefaultGraphHtpPrecision(value)[source]
Bases:
str
,Enum
An enumeration.
- FLOAT16 = 'FLOAT16'
- class OnnxOptions(execution_mode: Union[netspresso.np_qai.options.profile.ExecutionMode, NoneType] = <ExecutionMode.SEQUENTIAL: 'SEQUENTIAL'>, intra_op_num_threads: Union[int, NoneType] = 0, inter_op_num_threads: Union[int, NoneType] = 0, enable_memory_pattern: Union[bool, NoneType] = False, enable_cpu_memory_arena: Union[bool, NoneType] = False, graph_optimization_level: Union[netspresso.np_qai.options.profile.GraphOptimizationLevel, NoneType] = <GraphOptimizationLevel.ENABLE_ALL: 'ENABLE_ALL'>)[source]
Bases:
object
- execution_mode: ExecutionMode | None = 'SEQUENTIAL'
- intra_op_num_threads: int | None = 0
- inter_op_num_threads: int | None = 0
- enable_memory_pattern: bool | None = False
- enable_cpu_memory_arena: bool | None = False
- graph_optimization_level: GraphOptimizationLevel | None = 'ENABLE_ALL'
- class OnnxQnnOptions(execution_mode: Union[netspresso.np_qai.options.profile.ExecutionMode, NoneType] = <ExecutionMode.SEQUENTIAL: 'SEQUENTIAL'>, intra_op_num_threads: Union[int, NoneType] = 0, inter_op_num_threads: Union[int, NoneType] = 0, enable_memory_pattern: Union[bool, NoneType] = False, enable_cpu_memory_arena: Union[bool, NoneType] = False, graph_optimization_level: Union[netspresso.np_qai.options.profile.GraphOptimizationLevel, NoneType] = <GraphOptimizationLevel.ENABLE_ALL: 'ENABLE_ALL'>, qnn_htp_performance_mode: Union[netspresso.np_qai.options.profile.OnnxQnnHtpPerformanceMode, NoneType] = <OnnxQnnHtpPerformanceMode.BURST: 'burst'>, qnn_htp_graph_optimization_mode: Union[str, NoneType] = 3, qnn_enable_htp_fp16_precision: Union[str, NoneType] = 1)[source]
Bases:
OnnxOptions
- qnn_htp_performance_mode: OnnxQnnHtpPerformanceMode | None = 'burst'
- qnn_htp_graph_optimization_mode: str | None = 3
- qnn_enable_htp_fp16_precision: str | None = 1
- class TfliteOptions(enable_fallback: bool | NoneType = True, invoke_interpreter_on_cold_load: bool | NoneType = False, allow_fp32_as_fp16: bool | NoneType = True, force_opengl: bool | NoneType = False, number_of_threads: int | NoneType = -1, release_dynamic_tensors: bool | NoneType = False)[source]
Bases:
object
- enable_fallback: bool | None = True
- invoke_interpreter_on_cold_load: bool | None = False
- allow_fp32_as_fp16: bool | None = True
- force_opengl: bool | None = False
- number_of_threads: int | None = -1
- release_dynamic_tensors: bool | None = False
- class TfliteQnnOptions(enable_fallback: Union[bool, NoneType] = True, invoke_interpreter_on_cold_load: Union[bool, NoneType] = False, allow_fp32_as_fp16: Union[bool, NoneType] = True, force_opengl: Union[bool, NoneType] = False, number_of_threads: Union[int, NoneType] = -1, release_dynamic_tensors: Union[bool, NoneType] = False, qnn_log_level: Union[netspresso.np_qai.options.profile.QnnLogLevel, NoneType] = <QnnLogLevel.K_LOG_LEVEL_WARN: 'kLogLevelWarn'>, qnn_graph_priority: Union[netspresso.np_qai.options.profile.QnnGraphPriority, NoneType] = <QnnGraphPriority.K_QNN_PRIORITY_DEFAULT: 'kQnnPriorityDefault'>, qnn_gpu_precision: Union[netspresso.np_qai.options.profile.QnnGpuPrecision, NoneType] = <QnnGpuPrecision.K_GPU_FP16: 'kGpuFp16'>, qnn_gpu_performance_mode: Union[netspresso.np_qai.options.profile.QnnGpuPerformanceMode, NoneType] = <QnnGpuPerformanceMode.K_GPU_HIGH: 'kGpuHigh'>, qnn_dsp_performance_mode: Union[netspresso.np_qai.options.profile.QnnDspPerformanceMode, NoneType] = <QnnDspPerformanceMode.K_DSP_BURST: 'kDspBurst'>, qnn_dsp_encoding: Union[netspresso.np_qai.options.profile.QnnDspEncoding, NoneType] = <QnnDspEncoding.K_DSP_STATIC: 'kDspStatic'>, qnn_htp_performance_mode: Union[netspresso.np_qai.options.profile.TfliteQnnHtpPerformanceMode, NoneType] = <TfliteQnnHtpPerformanceMode.K_HTP_BURST: 'kHtpBurst'>, qnn_htp_precision: Union[netspresso.np_qai.options.profile.QnnHtpPrecision, NoneType] = <QnnHtpPrecision.K_HTP_FP16: 'kHtpFp16'>, qnn_htp_optimization_strategy: Union[netspresso.np_qai.options.profile.QnnHtpOptimizationStrategy, NoneType] = <QnnHtpOptimizationStrategy.K_HTP_OPTIMIZE_FOR_INFERENCE: 'kHtpOptimizeForInference'>, qnn_htp_use_conv_hmx: Union[bool, NoneType] = True, qnn_htp_use_fold_relu: Union[bool, NoneType] = False, qnn_htp_vtcm_size: Union[int, NoneType] = None, qnn_htp_num_hvx_threads: Union[int, NoneType] = None)[source]
Bases:
TfliteOptions
- qnn_log_level: QnnLogLevel | None = 'kLogLevelWarn'
- qnn_graph_priority: QnnGraphPriority | None = 'kQnnPriorityDefault'
- qnn_gpu_precision: QnnGpuPrecision | None = 'kGpuFp16'
- qnn_gpu_performance_mode: QnnGpuPerformanceMode | None = 'kGpuHigh'
- qnn_dsp_performance_mode: QnnDspPerformanceMode | None = 'kDspBurst'
- qnn_dsp_encoding: QnnDspEncoding | None = 'kDspStatic'
- qnn_htp_performance_mode: TfliteQnnHtpPerformanceMode | None = 'kHtpBurst'
- qnn_htp_precision: QnnHtpPrecision | None = 'kHtpFp16'
- qnn_htp_optimization_strategy: QnnHtpOptimizationStrategy | None = 'kHtpOptimizeForInference'
- qnn_htp_use_conv_hmx: bool | None = True
- qnn_htp_use_fold_relu: bool | None = False
- qnn_htp_vtcm_size: int | None = None
- qnn_htp_num_hvx_threads: int | None = None
- class TfliteGpuv2Options(enable_fallback: Union[bool, NoneType] = True, invoke_interpreter_on_cold_load: Union[bool, NoneType] = False, allow_fp32_as_fp16: Union[bool, NoneType] = True, force_opengl: Union[bool, NoneType] = False, number_of_threads: Union[int, NoneType] = -1, release_dynamic_tensors: Union[bool, NoneType] = False, gpu_inference_preference: Union[netspresso.np_qai.options.profile.GpuInferencePreference, NoneType] = <GpuInferencePreference.TFLITE_GPU_INFERENCE_PREFERENCE_SUSTAINED_SPEED: 'TFLITE_GPU_INFERENCE_PREFERENCE_SUSTAINED_SPEED'>, gpu_inference_priority1: Union[netspresso.np_qai.options.profile.GpuInferencePriority, NoneType] = <GpuInferencePriority.TFLITE_GPU_INFERENCE_PRIORITY_MIN_LATENCY: 'TFLITE_GPU_INFERENCE_PRIORITY_MIN_LATENCY'>, gpu_inference_priority2: Union[netspresso.np_qai.options.profile.GpuInferencePriority, NoneType] = <GpuInferencePriority.TFLITE_GPU_INFERENCE_PRIORITY_MIN_MEMORY_USAGE: 'TFLITE_GPU_INFERENCE_PRIORITY_MIN_MEMORY_USAGE'>, gpu_inference_priority3: Union[netspresso.np_qai.options.profile.GpuInferencePriority, NoneType] = <GpuInferencePriority.TFLITE_GPU_INFERENCE_PRIORITY_MAX_PRECISION: 'TFLITE_GPU_INFERENCE_PRIORITY_MAX_PRECISION'>, gpu_max_delegated_partitions: Union[int, NoneType] = 1)[source]
Bases:
TfliteOptions
- gpu_inference_preference: GpuInferencePreference | None = 'TFLITE_GPU_INFERENCE_PREFERENCE_SUSTAINED_SPEED'
- gpu_inference_priority1: GpuInferencePriority | None = 'TFLITE_GPU_INFERENCE_PRIORITY_MIN_LATENCY'
- gpu_inference_priority2: GpuInferencePriority | None = 'TFLITE_GPU_INFERENCE_PRIORITY_MIN_MEMORY_USAGE'
- gpu_inference_priority3: GpuInferencePriority | None = 'TFLITE_GPU_INFERENCE_PRIORITY_MAX_PRECISION'
- gpu_max_delegated_partitions: int | None = 1
- class TfliteNnapiOptions(enable_fallback: Union[bool, NoneType] = True, invoke_interpreter_on_cold_load: Union[bool, NoneType] = False, allow_fp32_as_fp16: Union[bool, NoneType] = True, force_opengl: Union[bool, NoneType] = False, number_of_threads: Union[int, NoneType] = -1, release_dynamic_tensors: Union[bool, NoneType] = False, nnapi_execution_preference: Union[netspresso.np_qai.options.profile.NnapiExecutionPreference, NoneType] = <NnapiExecutionPreference.K_SUSTAINED_SPEED: 'kSustainedSpeed'>, nnapi_max_number_delegated_partitions: Union[int, NoneType] = 3, nnapi_allow_fp16: Union[bool, NoneType] = True)[source]
Bases:
TfliteOptions
- nnapi_execution_preference: NnapiExecutionPreference | None = 'kSustainedSpeed'
- nnapi_max_number_delegated_partitions: int | None = 3
- nnapi_allow_fp16: bool | None = True
- class QnnOptions(default_graph_htp_optimization_value: Union[int, NoneType] = None, context_async_execution_queue_depth_numeric: Union[int, NoneType] = None, context_enable_graphs: Union[List[str], NoneType] = None, context_error_reporting_options_level: Union[netspresso.np_qai.options.profile.ContextErrorReportingOptionsLevel, NoneType] = None, context_error_reporting_options_storage_limit: Union[int, NoneType] = None, context_memory_limit_hint: Union[int, NoneType] = None, context_priority: Union[netspresso.np_qai.options.profile.Priority, NoneType] = None, context_gpu_performance_hint: Union[netspresso.np_qai.options.profile.ContextGpuPerformanceHint, NoneType] = <ContextGpuPerformanceHint.HIGH: 'HIGH'>, context_gpu_use_gl_buffers: Union[bool, NoneType] = None, context_htp_performance_mode: Union[netspresso.np_qai.options.profile.ContextHtpPerformanceMode, NoneType] = <ContextHtpPerformanceMode.BURST: 'BURST'>, default_graph_priority: Union[netspresso.np_qai.options.profile.Priority, NoneType] = True, default_graph_gpu_precision: Union[netspresso.np_qai.options.profile.DefaultGraphGpuPrecision, NoneType] = <DefaultGraphGpuPrecision.USER_PROVIDED: 'USER_PROVIDED'>, default_graph_gpu_disable_memory_optimizations: Union[bool, NoneType] = None, default_graph_gpu_disable_node_optimizations: Union[bool, NoneType] = None, default_graph_gpu_disable_queue_recording: Union[bool, NoneType] = None, default_graph_htp_disable_fold_relu_activation_into_conv: Union[bool, NoneType] = False, default_graph_htp_num_hvx_threads: Union[int, NoneType] = 4, default_graph_htp_optimization_type: Union[netspresso.np_qai.options.profile.DefaultGraphHtpOptimizationType, NoneType] = <DefaultGraphHtpOptimizationType.FINALIZE_OPTIMIZATION_FLAG: 'FINALIZE_OPTIMIZATION_FLAG'>, default_graph_htp_precision: Union[netspresso.np_qai.options.profile.DefaultGraphHtpPrecision, NoneType] = <DefaultGraphHtpPrecision.FLOAT16: 'FLOAT16'>, default_graph_htp_disable_short_depth_conv_on_hmx: Union[bool, NoneType] = False, default_graph_htp_vtcm_size: Union[int, NoneType] = 4)[source]
Bases:
object
- context_async_execution_queue_depth_numeric: int | None = None
- context_enable_graphs: List[str] | None = None
- context_error_reporting_options_level: ContextErrorReportingOptionsLevel | None = None
- context_error_reporting_options_storage_limit: int | None = None
- context_memory_limit_hint: int | None = None
- context_gpu_performance_hint: ContextGpuPerformanceHint | None = 'HIGH'
- context_gpu_use_gl_buffers: bool | None = None
- context_htp_performance_mode: ContextHtpPerformanceMode | None = 'BURST'
- default_graph_gpu_precision: DefaultGraphGpuPrecision | None = 'USER_PROVIDED'
- default_graph_gpu_disable_memory_optimizations: bool | None = None
- default_graph_gpu_disable_node_optimizations: bool | None = None
- default_graph_gpu_disable_queue_recording: bool | None = None
- default_graph_htp_disable_fold_relu_activation_into_conv: bool | None = False
- default_graph_htp_num_hvx_threads: int | None = 4
- default_graph_htp_optimization_type: DefaultGraphHtpOptimizationType | None = 'FINALIZE_OPTIMIZATION_FLAG'
- default_graph_htp_optimization_value: int | None = None
- default_graph_htp_precision: DefaultGraphHtpPrecision | None = 'FLOAT16'
- default_graph_htp_disable_short_depth_conv_on_hmx: bool | None = False
- default_graph_htp_vtcm_size: int | None = 4
- class ProfileCommonOptions(compute_unit: List[netspresso.np_qai.options.common.ComputeUnit] | NoneType = None, dequantize_outputs: bool | NoneType = True, tflite_delegates: List[netspresso.np_qai.options.profile.TfliteDelegates] | NoneType = None, tflite_options: netspresso.np_qai.options.profile.TfliteOptions | netspresso.np_qai.options.profile.TfliteQnnOptions | netspresso.np_qai.options.profile.TfliteGpuv2Options | netspresso.np_qai.options.profile.TfliteNnapiOptions | NoneType = None, qnn_options: netspresso.np_qai.options.profile.QnnOptions | NoneType = None, onnx_options: netspresso.np_qai.options.profile.OnnxOptions | netspresso.np_qai.options.profile.OnnxQnnOptions | NoneType = None, onnx_execution_providers: List[netspresso.np_qai.options.profile.OnnxExecutionProviders] | NoneType = None, max_profiler_iterations: int | NoneType = 100, max_profiler_time: int | NoneType = 600)[source]
Bases:
CommonOptions
- dequantize_outputs: bool | None = True
- tflite_delegates: List[TfliteDelegates] | None = None
- tflite_options: TfliteOptions | TfliteQnnOptions | TfliteGpuv2Options | TfliteNnapiOptions | None = None
- qnn_options: QnnOptions | None = None
- onnx_options: OnnxOptions | OnnxQnnOptions | None = None
- onnx_execution_providers: List[OnnxExecutionProviders] | None = None
- max_profiler_iterations: int | None = 100
- max_profiler_time: int | None = 600
- class ProfileOptions(compute_unit: List[ComputeUnit] | None = None, dequantize_outputs: bool | None = True, tflite_delegates: List[TfliteDelegates] | None = None, tflite_options: TfliteOptions | TfliteQnnOptions | TfliteGpuv2Options | TfliteNnapiOptions | None = None, qnn_options: QnnOptions | None = None, onnx_options: OnnxOptions | OnnxQnnOptions | None = None, onnx_execution_providers: List[OnnxExecutionProviders] | None = None, max_profiler_iterations: int | None = 100, max_profiler_time: int | None = 600)[source]
Bases:
ProfileCommonOptions
Profile options for the model.
Note
For details, see ProfileOptions in QAI Hub API.
- class InferenceOptions(compute_unit: List[ComputeUnit] | None = None, dequantize_outputs: bool | None = True, tflite_delegates: List[TfliteDelegates] | None = None, tflite_options: TfliteOptions | TfliteQnnOptions | TfliteGpuv2Options | TfliteNnapiOptions | None = None, qnn_options: QnnOptions | None = None, onnx_options: OnnxOptions | OnnxQnnOptions | None = None, onnx_execution_providers: List[OnnxExecutionProviders] | None = None, max_profiler_iterations: int | None = 100, max_profiler_time: int | None = 600)[source]
Bases:
ProfileCommonOptions
Inference options for the model.
Note
For details, see InferenceOptions in QAI Hub API.