Skip to content

Add algorithm search to FP8 sparse linear benchmark (#4397)#4397

Open
bbeckca wants to merge 1 commit into
pytorch:mainfrom
bbeckca:export-D102683062
Open

Add algorithm search to FP8 sparse linear benchmark (#4397)#4397
bbeckca wants to merge 1 commit into
pytorch:mainfrom
bbeckca:export-D102683062

Conversation

@bbeckca
Copy link
Copy Markdown
Contributor

@bbeckca bbeckca commented May 13, 2026

Summary:

What: Adds algorithm search support (--search-alg) to the FP8 sparse linear benchmark. Threads a new alg_id parameter through the quantization config (Float8DynamicActivationFloat8WeightConfig), the sparse tensor class, and down into the _cslt_sparse_mm kernel call. When --search-alg is passed, the benchmark calls _cslt_sparse_mm_search to find the best algorithm for each shape and benchmarks with that algorithm to report the speedup.

Why: hipSPARSELt supports multiple algorithms for sparse matmul, and the default (alg_id=0) isn't always the fastest. This lets us find the best algorithm for a given shape and measure the performance benefit.

Differential Revision: D102683062

@pytorch-bot
Copy link
Copy Markdown

pytorch-bot Bot commented May 13, 2026

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/ao/4397

Note: Links to docs will display an error until the docs builds have been completed.

❗ 1 Active SEVs

There are 1 currently active SEVs. If your PR is affected, please view them below:

❌ 2 New Failures

As of commit 926a428 with merge base 13cd013 (image):

NEW FAILURES - The following jobs have failed:

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@meta-cla meta-cla Bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label May 13, 2026
@meta-codesync
Copy link
Copy Markdown

meta-codesync Bot commented May 13, 2026

@bbeckca has exported this pull request. If you are a Meta employee, you can view the originating Diff in D102683062.

@bbeckca
Copy link
Copy Markdown
Contributor Author

bbeckca commented May 14, 2026

@pytorchbot label "module: rocm"

@meta-codesync meta-codesync Bot changed the title Add algorithm search to FP8 sparse linear benchmark Add algorithm search to FP8 sparse linear benchmark (#4397) May 14, 2026
@bbeckca bbeckca force-pushed the export-D102683062 branch from 26fe93b to b918f9b Compare May 14, 2026 18:05
bbeckca added a commit to bbeckca/ao that referenced this pull request May 14, 2026
Summary:

What: Adds algorithm search support (--search-alg) to the FP8 sparse linear benchmark. Threads a new alg_id parameter through the quantization config (Float8DynamicActivationFloat8WeightConfig), the sparse tensor class, and down into the _cslt_sparse_mm kernel call. When --search-alg is passed, the benchmark calls _cslt_sparse_mm_search to find the best algorithm for each shape and benchmarks with that algorithm to report the speedup.

Why: hipSPARSELt supports multiple algorithms for sparse matmul, and the default (alg_id=0) isn't always the fastest. This lets us find the best algorithm for a given shape and measure the performance benefit.

Differential Revision: D102683062
bbeckca added a commit to bbeckca/ao that referenced this pull request May 14, 2026
Summary:

What: Adds algorithm search support (--search-alg) to the FP8 sparse linear benchmark. Threads a new alg_id parameter through the quantization config (Float8DynamicActivationFloat8WeightConfig), the sparse tensor class, and down into the _cslt_sparse_mm kernel call. When --search-alg is passed, the benchmark calls _cslt_sparse_mm_search to find the best algorithm for each shape and benchmarks with that algorithm to report the speedup.

Why: hipSPARSELt supports multiple algorithms for sparse matmul, and the default (alg_id=0) isn't always the fastest. This lets us find the best algorithm for a given shape and measure the performance benefit.

Differential Revision: D102683062
@bbeckca bbeckca force-pushed the export-D102683062 branch from b918f9b to 321ecbb Compare May 14, 2026 18:15
Summary:

What: Adds algorithm search support (--search-alg) to the FP8 sparse linear benchmark. Threads a new alg_id parameter through the quantization config (Float8DynamicActivationFloat8WeightConfig), the sparse tensor class, and down into the _cslt_sparse_mm kernel call. When --search-alg is passed, the benchmark calls _cslt_sparse_mm_search to find the best algorithm for each shape and benchmarks with that algorithm to report the speedup.

Why: hipSPARSELt supports multiple algorithms for sparse matmul, and the default (alg_id=0) isn't always the fastest. This lets us find the best algorithm for a given shape and measure the performance benefit.

Differential Revision: D102683062
@bbeckca bbeckca force-pushed the export-D102683062 branch from 321ecbb to 926a428 Compare May 15, 2026 20:37
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. fb-exported meta-exported module: rocm

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant