Skip to content

Conversation

@GunnarFarneback
Copy link
Collaborator

This PR adds the possibility to specify provider options in load_inference.

Of particular interest is the CUDA provider option cudnn_conv_algo_search. The default EXHAUSTIVE option is, although fastest, numerically somewhat unreliable. On newer GPUs it can cause Float32 convolution layers to produce results which vary with like 1000 eps between different sessions. For certain applications it's therefore valuable to be able to choose the (at least empirically) more stable DEFAULT and HEURISTIC options.

Copy link
Owner

@jw3126 jw3126 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM thanks!

@jw3126 jw3126 merged commit 11e6c3b into jw3126:main Apr 13, 2024
@GunnarFarneback GunnarFarneback deleted the cuda_provider_options branch April 13, 2024 16:24
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants