Skip to content

Tags: jnothman/pytorch

Tags

ciflow/trunk/82290

Toggle ciflow/trunk/82290's commit message

Verified

This commit was created on GitHub.com and signed with GitHub’s verified signature. The key has expired.
Remove trailing spaces

ciflow/trunk/82279

Toggle ciflow/trunk/82279's commit message

Verified

This commit was created on GitHub.com and signed with GitHub’s verified signature. The key has expired.
Resolve issue 74968 TODO

As pytorch#74968 is closed now.

ciflow/trunk/82028

Toggle ciflow/trunk/82028's commit message
[PyTorch][Kineto] add ActivityType.h when USE_KINETO is not set (pyto…

…rch#82028)

Summary:
Pull Request resolved: pytorch#82028

This patch fixes an error "'ActivityType.h' file not found" when `use_kineto()` is false.

## Problem
Even when `use_kineto()` is not set (i.e., `-DUSE_KINETO` is not passed), `ActivityType.h` is required for PyTorch compilation:
https://github.com/pytorch/pytorch/blob/master/torch/csrc/profiler/kineto_shim.h#L15

## Solution
Add `ActivitiyType.h` dependency even when `use_kineto() == False`.

Test Plan: PyTorch internal and external CI tests.

Differential Revision: D38090153

fbshipit-source-id: 73fad78a3d3576a060850012cc9444ac070398d3

ciflow/trunk/81946

Toggle ciflow/trunk/81946's commit message
[pytorch] Bump SoLoader version to 0.10.4 (pytorch#81946)

Summary: Pull Request resolved: pytorch#81946

Test Plan: sandcastle CI.

Reviewed By: linbinyu

Differential Revision: D38056789

fbshipit-source-id: 9f2d70c5bdc6e90b084bea5f2da59da531931871

ciflow/trunk/81861

Toggle ciflow/trunk/81861's commit message
Update on "[NVFuser] Upstream push 0714"


Syncing nvfuser devel branch to upstream master. https://github.com/csarofeen/pytorch/

Code changes includes:

- codegen improvements:
  1. Indexing refactor -> Remove reference tensor in predicate indexing logic
  2. MMA Rfactor support for cross-warp and cross-CTA split on K dimension
  3. Grouping grid allreduces across iterations
  4. Swizzle op formulation for non-affine swizzles
  5. Use scheduler_utils to cache inputs and outputs in schedulePointwise
- scheduler refactor
  1. New compute at interface
- transformation propagation refactor on MaxInfoSpanningTree
  1. Added sibling path that is required to generate consistent replay for some cases where `MaxInfoSpanningTree` is used with a selector.
  2. Optimization to skip Transform propagator
  3. SpanningTreePrinter for debugging
- parser update
  1. Fixes `div`
  2. Added `_to_copy`
  3. Broadcast in dim with expand to support expanding to concrete size
  4. Dropout prob extremal patch
- executor patch on caching strides for output allocation

Squashed commits to WAR github API
Commits that's actually in this PR from the devel branch:

```
3b87896 Fix allocation of work buffers and `fused_reduction::ParallelReduce` with unswitch (pytorch#1818)
4cae122 schedulePointwise cleanup: - computeAt + InlinePropagator (pytorch#1815)
3df9742 Use scheduler_utils to cache inputs and outputs in schedulePointwise (pytorch#1811)
03180aa improve broadcast resolution (pytorch#1792)
bee6c69 bug fix (pytorch#1819)
4413c8f Support PYTORCH_NVFUSER_DUMP=transform_propagator (pytorch#1812)
de6b7ca Fix negative position in InlinePropagator (pytorch#1813)
10a996c Remove redundant check in schedulePointwise (pytorch#1810)
acd5ed4 Swizzle op formulation for non-affine swizzles (pytorch#1441)
3ed8330 Kernel args patch to show zero_init buffer (pytorch#1809)
037a75a Dropout prob extremal patch (pytorch#1804)
282c429 spam nvrtc options (pytorch#1783)
3ba6a5f Broadcast in dim with expand (pytorch#1794)
fd4be12 remove dead indexing code (pytorch#1806)
fa4e6a4 Check siblings in getMaxPosAll (pytorch#1805)
025c840 Grouping grid allreduces across iterations (pytorch#1755)
37c579e Temporarily disable test requring large shared memory. (pytorch#1802)
5f375d0 More cleanup on InlinePropagator (pytorch#1800)
8d384da Indexing refactor stage 2 : Remove reference tensor in predicate indexing logic (pytorch#1784)
f008140 MMA Rfactor support for cross-warp and cross-CTA split on K dimension (pytorch#1554)
76b3cca Add parsing support for `_to_copy` to handle AMP casts. (pytorch#1756)
ef04f6c Coding style cleanups (pytorch#1798)
38c7f3c InlinePropagator please don't replay (pytorch#1797)
3f2c263 validateDomain in TransformPropagator (pytorch#1796)
c077085 Use TransformPropagatorWithCheck in many tests (pytorch#1795)
d0d0908 Some further cleanup for the new computeAt interface (pytorch#1793)
45f5203 Fix TransformReplay::getMatchedLeafPosWithoutReplay* (pytorch#1791)
28cbaf9 New compute at interface (pytorch#1743)
635ebfc Add SpanningTreePrinter (pytorch#1786)
59f3c32 Output allocate patch (pytorch#1790)
fe93bf5 Transform propagator skip replay when possible (pytorch#1782)
ebf23a5 Fix isIntegralType error msg (pytorch#1789)
0c82ecf Disable register reuse across serial broadcast ops (pytorch#1787)
33a824d Adding sibling path for MaxInfoSpanningTree (pytorch#1776)
86f46aa Fix div(Val, TensorView) (pytorch#1778)
d3de227 Fix FusionMaxRootDomainInfoSpanningTreePrintTwice_CUDA (pytorch#1781)
ecc7a87 Extend mma dimension and layout checking to support strided batched matmul and tensor contractions (pytorch#1761)
```

RUN_TORCHBENCH: nvfuser

Differential Revision: [D38043938](https://our.internmc.facebook.com/intern/diff/D38043938)

[ghstack-poisoned]

ciflow/trunk/79975

Toggle ciflow/trunk/79975's commit message
more tests :////

ciflow/periodic/82291

Toggle ciflow/periodic/82291's commit message
Update on "[_shard] Add ShardedTensorBase"

This PR added ShardedTensorBase, which is the base class of
ShardedTensor, and only contains local shards, ShardedTensorMetadata,
and does not have any communication backend attached (i.e ProcessGroup)

Differential Revision: [D38190272](https://our.internmc.facebook.com/intern/diff/D38190272)

[ghstack-poisoned]

ciflow/periodic/82290

Toggle ciflow/periodic/82290's commit message

Verified

This commit was created on GitHub.com and signed with GitHub’s verified signature. The key has expired.
Remove trailing spaces

ciflow/periodic/82279

Toggle ciflow/periodic/82279's commit message

Verified

This commit was created on GitHub.com and signed with GitHub’s verified signature. The key has expired.
Resolve issue 74968 TODO

As pytorch#74968 is closed now.

ciflow/periodic/82028

Toggle ciflow/periodic/82028's commit message
[PyTorch][Kineto] add ActivityType.h when USE_KINETO is not set (pyto…

…rch#82028)

Summary:
Pull Request resolved: pytorch#82028

This patch fixes an error "'ActivityType.h' file not found" when `use_kineto()` is false.

## Problem
Even when `use_kineto()` is not set (i.e., `-DUSE_KINETO` is not passed), `ActivityType.h` is required for PyTorch compilation:
https://github.com/pytorch/pytorch/blob/master/torch/csrc/profiler/kineto_shim.h#L15

## Solution
Add `ActivitiyType.h` dependency even when `use_kineto() == False`.

Test Plan: PyTorch internal and external CI tests.

Differential Revision: D38090153

fbshipit-source-id: 73fad78a3d3576a060850012cc9444ac070398d3