Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
1586 commits
Select commit Hold shift + click to select a range
d0c3297
Refine XPU allocator message when OOM (#165509)
guangyey Oct 15, 2025
eaeaa08
[PowerPC] Disable MKLDNN TF32 on PowerPC to fix build failure (#163454)
Tiwari-Avanish Oct 16, 2025
d73c283
[CUDA] Large tensor maxpool crash fix (#165374)
Isalia20 Oct 16, 2025
69b0591
Revert "Add mingw to docker (#165560)"
pytorchmergebot Oct 16, 2025
f06e669
refactor: replace runtime_error with TORCH_CHECK for better error han…
licy666 Oct 16, 2025
9272437
Fx collectives bucketing: add bucket all_reduce (#165351)
IvanKobzarev Oct 16, 2025
e6033f6
[MPS] Improve `index_fill_` error handling (#165594)
malfet Oct 16, 2025
8573574
[MPS] sparse mask implementation (#165102)
Isalia20 Oct 16, 2025
1a5b7ec
[BE] Fold cond into `TORCH_CHECK(false,...)` (#165593)
malfet Oct 16, 2025
e6d9d68
[Bugfix][Dynamo] Fix Sparse tensors by graph break in Dynamo (#164873)
Lucaskabela Oct 16, 2025
7ee45f7
Restore AcceleratorAllocatorConfig to avoid potential regression (#16…
guangyey Oct 14, 2025
03e5dbb
Register CUDAAllocatorConfig to AcceleratorAllocatorConfig (#165131)
guangyey Oct 14, 2025
608a6d4
Reuse AcceleratorAllocatorConfig in CUDAAllocatorConfig (#165135)
guangyey Oct 14, 2025
515b5ff
Remove unused code in CUDAAllocatorConfig (#165136)
guangyey Oct 14, 2025
219fb6a
Refactor CUDAAllocatorConfig using ConfigTokenizer (#165281)
guangyey Oct 14, 2025
f33c7e1
add and fix OpInfo tests for the default partitioner (#165372)
bdhirsh Oct 16, 2025
ed74dc0
add the option to disable functionalization in AOTDispatcher (#164577)
bdhirsh Oct 16, 2025
783da8b
Repro for property related Dynamo graph break (#165609)
ezyang Oct 16, 2025
99b32a6
[inductor] print 0.0 as 0 for triton (#164291)
isuruf Oct 15, 2025
d61a9b8
[DeviceMesh] Prefer using _layout over _mesh for all sorts of things …
lw Oct 16, 2025
e1d71a6
Revert "12/n : Remove fbandroid_compiler_flags (#165558)"
pytorchmergebot Oct 16, 2025
85586d7
Make c7i the default for _linux-build.yml (#164747)
zxiiro Oct 16, 2025
fe5ccb1
bf16 support for per tensor backward (#165362)
liangel-02 Oct 14, 2025
1a34ff4
Fixing get_local_rank() variable missing when compiled (#165432)
arkadip-maitra Oct 16, 2025
7d87d70
[inductor][bucketing] Fx collectives bucketing of multiple dtypes (#1…
IvanKobzarev Oct 16, 2025
a214371
[FP8] Add other Blackwell compute-capabiilities to expected fail `tes…
eqy Oct 16, 2025
99097b6
[DeviceMesh] Introduce private constructor instead of _create_mesh_fr…
lw Oct 16, 2025
86fd4fc
[DeviceMesh] Simplify unflatten method (#165556)
lw Oct 16, 2025
7669ac9
[ROCm] Add scaled_mm v2 support. (#165528)
jagadish-amd Oct 16, 2025
a303d6d
[inductor] don't try to reorder loops for template (#165601)
shunting314 Oct 16, 2025
6dedd34
[CD] Skip 12.9 build on Windows (#165665)
huydhn Oct 16, 2025
5daef30
158232 Fix autocast cache incorrectly retaining no_grad state (#165068)
stmcgovern Oct 16, 2025
d4a713c
Change forkserver test to only run below 3.13.8 (#165667)
yushangdi Oct 16, 2025
7df9aca
[ROCm][Windows] Enable AOTriton runtime compile on Windows (#165538)
tvukovic-amd Oct 16, 2025
d795fb2
[RFC] Add pyrefly to lintrunner (#165179)
maggiemoss Oct 16, 2025
585b9db
[async_tp] Support ag+mm with gather_dim lastdim of mat_A (#163068)
IvanKobzarev Oct 15, 2025
37f3ba2
[Fix] Use sys.executable instead of hardcoded python (#165633)
lingebeng Oct 16, 2025
aba8c43
Register var for MTIA (#165382)
trirpi Oct 16, 2025
9bf5b38
[Inductor][Triton][FP8] Refactor scaled_mm template to accept scaling…
jananisriram Oct 16, 2025
aead927
12/n : Remove fbandroid_compiler_flags (#165558)
kambati-meta Oct 16, 2025
431c13c
Revert "[DeviceMesh] Simplify unflatten method (#165556)"
pytorchmergebot Oct 16, 2025
b10f463
Revert "[DeviceMesh] Introduce private constructor instead of _create…
pytorchmergebot Oct 16, 2025
27a98e6
Revert "[DeviceMesh] Prefer using _layout over _mesh for all sorts of…
pytorchmergebot Oct 16, 2025
fb06e49
Revert "[inductor] print 0.0 as 0 for triton (#164291)"
pytorchmergebot Oct 16, 2025
7d0f872
Use union syntax in torch/_inductor runtime and fx_passes (#165652)
oulgen Oct 16, 2025
2cd5fd1
Enable local tensor mode on DTensor view ops test (#165596)
dzmitry-huba Oct 16, 2025
e86942f
minor proxy_tensor reorg (#165266)
aorenste Oct 12, 2025
5f21cc7
Teach ProxyTorchDispatchMode how to decompose sympy.Expr into known i…
aorenste Oct 12, 2025
4c1c341
FakeTensorMode shouldn't cache syms when tracing (#164718)
aorenste Oct 15, 2025
1a54d33
[easy] Fix graph_capture in aot_joint_with_descriptors test (#165660)
yiming0416 Oct 16, 2025
cbc08c8
Add NEON acceleration for `Vectorized<int[8|16|32|64>` (#165273)
Nicoshev Oct 16, 2025
5641de7
Add suppressions for _inductor/codegen (#165659)
maggiemoss Oct 16, 2025
d5db3ae
[CI] Use 1-GPU runners for rocm-mi355.yml (#165658)
jithunnair-amd Oct 16, 2025
d7e275d
[CI][CUDA] Add periodic b200 distributed job (#159323)
nWEIdia Oct 16, 2025
4d833f8
[BE] [CI] Fix aarch64 arch checks (#165676)
malfet Oct 16, 2025
ce109b3
Add `torch.backends.mkldnn.is_acl_available()` method (#165678)
malfet Oct 16, 2025
556fc09
[DebugMode][1/N] refactor logs into _DebugCalls (#165376)
pianpwk Oct 13, 2025
5b3ea75
[Mem Snapshot] Add Metadata Field (#165490)
sraikund16 Oct 16, 2025
98a488c
Start recording inductor provenance (#162669)
c00w Oct 16, 2025
d2c82ba
Revert "158232 Fix autocast cache incorrectly retaining no_grad stat…
pytorchmergebot Oct 16, 2025
e0fe37f
[MPS] Move `torch.cat` impl to Metal (#165373)
kurtamohler Oct 14, 2025
470e2f6
Revert "[Fix] Use sys.executable instead of hardcoded python (#165633)"
pytorchmergebot Oct 17, 2025
b2953f5
[9/N] Apply ruff UP035 rule (#165515)
cyyever Oct 17, 2025
5b2afe4
Turn some const variables into constexpr in C++ code (#165401)
cyyever Oct 17, 2025
5d9b024
Add mingw to docker (#165560)
yushangdi Oct 16, 2025
d82527b
[Windows] Add AOTI cross-compilation CI (#165573)
yushangdi Oct 16, 2025
9726553
[BE][Ez]: Use sys.executable instead of hardcoded Python (#165679)
Skylion007 Oct 17, 2025
11e2084
Revert "[Mem Snapshot] Add Metadata Field (#165490)"
pytorchmergebot Oct 17, 2025
d0add0b
[torchfuzz] check in some more ignore regexes (#164749)
bobrenjc93 Oct 16, 2025
7dabfb0
[torchfuzz] add support for --stop-at-first-failure flag (#165529)
bobrenjc93 Oct 16, 2025
9fccbdd
Fix incorrect function signature in template (#165567)
muchulee8 Oct 15, 2025
3154482
[CUDA][cuBLAS] Only `xFail` `addmm` with reduced precision reductions…
eqy Oct 17, 2025
861cdb8
use statically_known_leq & *=2 instead of bound_sympy in persistent r…
eellison Oct 16, 2025
fcbde24
[ONNX] Remove common imports from torchlib (#165156)
justinchuby Oct 17, 2025
43d7842
Pyrefly suppressions 2 (#165692)
maggiemoss Oct 17, 2025
7e15046
allow providing full fr trace path (#165639)
tushar00jain Oct 17, 2025
364624e
[codemod][lowrisk] Remove unused exception parameter from some files …
r-barnes Oct 17, 2025
9e94ec7
Revert "Turn some const variables into constexpr in C++ code (#165401)"
pytorchmergebot Oct 17, 2025
24879f0
[dynamo] Use Variable Builder to build the property fget object (#165…
anijain2305 Oct 16, 2025
f1d8822
[annotate] add annotate_fn function decorator (#165703)
yushangdi Oct 17, 2025
e925dfc
Enable all SIM rules except disabled ones (#164645)
cyyever Oct 17, 2025
fdd560a
[export] preserve_node_meta by default (#165524)
pianpwk Oct 17, 2025
51348c0
Give a friendly message for older Intel GPU (#165622)
guangyey Oct 17, 2025
b44fb14
Remove unused parameter when query extension attribute (#165623)
guangyey Oct 17, 2025
d0c24b3
[APF Logging][Error Trait] To fill the errorTraits for ChildFailedErr…
XilunWu Oct 16, 2025
9fe3b2a
Remove torch.serialization entries from the doc ignore list (#160224)
jbschlosser Aug 8, 2025
202f83d
[ROCm][layer_norm] Use __builtin_amdgcn_rcpf(x) instead of 1.f/x (#16…
jerrymannil Oct 17, 2025
cb6e4d7
User-passed alpha to scaled_gemm (#165563)
slayton58 Oct 16, 2025
4a22139
[MPS][BE] Fix unused variable warning (#165726)
malfet Oct 17, 2025
80d2ca7
Revert "[annotate] add annotate_fn function decorator (#165703)"
pytorchmergebot Oct 17, 2025
574c9fc
Revert "Remove torch.serialization entries from the doc ignore list (…
pytorchmergebot Oct 17, 2025
5d4da26
Revert "[export] preserve_node_meta by default (#165524)"
pytorchmergebot Oct 17, 2025
7231118
Turn some const variables into constexpr in C++ code (#165401)
cyyever Oct 17, 2025
ce29d0d
[ATen] Vectorize 8 elements on 16 bit data types for sum/mean (#165055)
PaulZhang12 Oct 16, 2025
6ece527
[CI] Add aarch64 operator benchmark (#165585)
malfet Oct 16, 2025
3af2f0c
[inductor] require shape in TritonCSEVariable (#162275)
isuruf Sep 30, 2025
935ccdb
[MPS] Fix internal assertion in torch.linalg.solve for singular matri…
inventshah Oct 17, 2025
85c5433
Revert "Fix `_StridedShard` incorrect split (#165533)"
pytorchmergebot Oct 17, 2025
faff826
Revert "[ROCm] new implementation of upsample_bilinear2d_backward (#1…
pytorchmergebot Oct 17, 2025
bfcdbd0
fix wrong accuracy_status when exception. (#165731)
xuhancn Oct 17, 2025
1dc9a05
[dynamo][user_defined] Replace UserFunctionVariable with VariableTrac…
anijain2305 Oct 17, 2025
630520b
[dynamo][misc] Replace UserFunctionVariable with VariableTracker buil…
anijain2305 Oct 17, 2025
2928c5c
Revert "Pyrefly suppressions 2 (#165692)"
pytorchmergebot Oct 17, 2025
080365b
Escaped html tags name and target to appear as strings (#165543)
trichmo Oct 17, 2025
45afaf0
[DebugMode][2/N] add nn.Module tracking (#165498)
pianpwk Oct 17, 2025
da8517f
[ROCm][CI] upgrade wheels to 7.0.2 and 6.4.4 patch release (#165756)
jeffdaily Oct 17, 2025
cff1b20
Patch the flex_attention._get_mod_type to not use inspect.signature w…
jmaczan Oct 17, 2025
dd3b48e
Fix bug with serialization after AOTAutogradCache hit (#165474)
jamesjwu Oct 14, 2025
39e0a83
Fix B200 test fails in scaled_mm (#165747)
slayton58 Oct 17, 2025
a032510
shrink_group implementation to expose ncclCommShrink API (#164518)
brchang24 Oct 17, 2025
58879bf
[DeviceMesh] Prefer using _layout over _mesh for all sorts of things …
lw Oct 17, 2025
d659bbd
[DeviceMesh] Introduce private constructor instead of _create_mesh_fr…
lw Oct 17, 2025
0d4c2b7
[DeviceMesh] Simplify unflatten method (#165556)
lw Oct 17, 2025
9a71d96
Revert "[DebugMode][1/N] refactor logs into _DebugCalls (#165376)"
pytorchmergebot Oct 17, 2025
ca5b7f8
torch.compile: populate compiler_config (#165581)
c00w Oct 17, 2025
b08d8c2
Revert "[DebugMode][2/N] add nn.Module tracking (#165498)"
pytorchmergebot Oct 17, 2025
3806e97
Refactor out headeronly ArrayRef (#164991)
janeyx99 Oct 16, 2025
e445494
Widen ops support to take in IntHOArrayRef vs only std::vec (#165152)
janeyx99 Oct 16, 2025
7a65770
Update gm.print_readable to include Annotation (#165397)
SherlockNoMad Oct 17, 2025
fae74cd
Revert "shrink_group implementation to expose ncclCommShrink API (#16…
pytorchmergebot Oct 17, 2025
08c97b4
Don't run compile inside kernel invocation (#165687)
tugsbayasgalan Oct 17, 2025
9c12651
Improve error message for non-positive groups in convolution (#165669)
vishalgoyal316 Oct 17, 2025
a664b29
Update docs for torch.mode (#165614)
kasparas-k Oct 17, 2025
382b015
[docs] Add usage examples to ConvTranspose1d docstring (#165618)
vishalgoyal316 Oct 17, 2025
a16fd6b
[NVSHMEM][Triton] Fix NVSHMEM triton test for wacky world sizes (#165…
eqy Oct 17, 2025
75e2a9f
[annotate] add annotate_fn function decorator (#165703)
yushangdi Oct 17, 2025
2bcd892
[distributed] Replace assert statements in distributed checkpoint wit…
RohitRathore1 Oct 17, 2025
6c9c6e0
Enable C407 of flake8 (#165046)
cyyever Oct 17, 2025
06d3243
Revert "Escaped html tags name and target to appear as strings (#1655…
pytorchmergebot Oct 17, 2025
ab65498
Fix `_StridedShard` incorrect split (#165533)
zpcore Oct 17, 2025
8cb2fb4
[Inductor] Support fallback for all gemm like ops (#165755)
nandesuka Oct 17, 2025
86ebce1
[precompile] Pass tensor_to_context to backend. (#165702)
zhxchen17 Oct 17, 2025
c18ddfc
[dynamo][easy] Support torch.accelerator.current_accelerator (#165734)
anijain2305 Oct 17, 2025
616c6bd
[dynamo][ac] Config flag to allow eager and compile AC divergence for…
anijain2305 Oct 17, 2025
2e22b1a
[pytorch] Composite backend potential fix for is_backend_available (#…
wesbland Oct 17, 2025
e50dc40
Revert "Update gm.print_readable to include Annotation (#165397)"
pytorchmergebot Oct 17, 2025
fe80f03
Add B200 files to labeler and update codeowners (#165767)
drisspg Oct 17, 2025
1b39742
Enable more DTensor tests in local tensor mode and fix more integrati…
dzmitry-huba Oct 17, 2025
69c3389
Revert "[Inductor][CuTeDSL] Move load_template up two directories (#1…
pytorchmergebot Oct 17, 2025
a25a649
[Mem Snapshot] Add Metadata Field (#165490)
sraikund16 Oct 17, 2025
29b0296
Fixed issue with GradTrackingTensor not properly propagating sparse l…
cleonard530 Oct 18, 2025
e9f4999
[Code Clean] Replace std::runtime_error with TORCH_CHECK (#165305)
orangeH25 Oct 18, 2025
543ddbf
[ONNX] Support renaming in dynamic axes to shapes conversion (#165769)
titaiwangms Oct 18, 2025
de3da77
Thread deterministic config vars to subproc compilation (#165729)
drisspg Oct 17, 2025
cf3a787
[annotate] Annotate bw nodes before eliminate dead code (#165782)
yushangdi Oct 18, 2025
c137e22
.venv/ in .gitignore (#165418)
jmaczan Oct 18, 2025
de09bab
[BE]: Update cudnn frontend submodule to 1.15.0 (#165776)
Skylion007 Oct 18, 2025
c6a8db0
Fix issues with generalized_scatter and setitem allocated unbacked s…
laithsakka Oct 17, 2025
017d298
set unbacked bindings in reinplace pass for newly created nodes durin…
laithsakka Oct 17, 2025
e4d6c56
Improve dynamo graph capture stack trace for custom ops (#165693)
yiming0416 Oct 18, 2025
23417ae
[Submodule] Bump FBGEMM to latest (#165544)
slayton58 Oct 17, 2025
d9f94e0
[dynamo] Support fx.traceback.annotate as decorator (#165805)
anijain2305 Oct 17, 2025
9095a9d
[CD] Apply the fix from #162455 to aarch64+cu129 build (#165794)
huydhn Oct 18, 2025
f02e394
Expand type checking to mypy strict files (#165697)
maggiemoss Oct 18, 2025
b819426
Remove unnecessary noqa suppressions (#164106)
cyyever Oct 18, 2025
0f0b4bf
[1/N] Remove unused header inclusion (#165763)
cyyever Oct 18, 2025
aaac8cb
[1/N] Add strict parameter to Python zip calls (#165531)
cyyever Oct 18, 2025
e595136
Enable PLC1802 on ruff (#165813)
cyyever Oct 18, 2025
c79dfdc
Enable all PIE rules on ruff (#165814)
cyyever Oct 18, 2025
24520b8
Revert "Enable all PIE rules on ruff (#165814)"
pytorchmergebot Oct 18, 2025
0bbdd6b
[ROCm][inductor] heuristic improvements for pointwise kernels (#163197)
naromero77amd Oct 18, 2025
a0948d4
[ROCm][inductor] autotune support for persistent reduction kernels (#…
naromero77amd Oct 18, 2025
fdab48a
Enable all PIE rules on ruff (#165814)
cyyever Oct 18, 2025
ad67170
[MPS] sparse matmuls (#165232)
Isalia20 Oct 18, 2025
4740ce7
[CP] Fix load balancer incorrectly assuming batch dimension exists (#…
fegin Oct 18, 2025
beb6b62
Revert "Enable more DTensor tests in local tensor mode and fix more i…
pytorchmergebot Oct 18, 2025
f510d0d
Clarrifying input output angle unit in the docs for trigonometric fun…
arkadip-maitra Oct 18, 2025
d14cbb4
Add NVFP4 two-level scaling to scaled_mm (#165774)
slayton58 Oct 17, 2025
032bed9
Various C++ code fixes in LSAN integration (#165818)
cyyever Oct 18, 2025
1f43d17
Fix self assignment (#165816)
cyyever Oct 18, 2025
35e5189
Remove CUDA 11 workarounds for CUB_SUPPORTS_SCAN_BY_KEY and CUB_SUPPO…
cyyever Oct 18, 2025
f18041c
Fix missing closing quote in __init__.py documentation (#165827)
andreh7 Oct 18, 2025
c4f6619
Enable more DTensor tests in local tensor mode and fix more integrati…
dzmitry-huba Oct 18, 2025
3255e78
Enable all flake8-logging-format rules (#164655)
cyyever Oct 19, 2025
e939651
[audio hash update] update the pinned audio hash (#165807)
pytorchupdatebot Oct 19, 2025
33adb27
[BE][Ez]: Update Eigen to 5.0.0. C++14 support and more! (#165840)
Skylion007 Oct 19, 2025
ceb11a5
[BE]: Update kleidai submodule to v1.15.0 (#165842)
Skylion007 Oct 19, 2025
57ba575
[BE][Ez]: Update torch.is_tensor documentation (#165841)
Skylion007 Oct 19, 2025
5d62b63
[BE] Use Python-3.14 GE build (#165804)
malfet Oct 17, 2025
4888ed4
Refine Allocator Config error message friendly (#165288)
guangyey Oct 17, 2025
a1114be
Deprecate overlapped functions in CUDAAllocatorConfig (#165289)
guangyey Oct 17, 2025
b2f5c25
Introduce a generic API torch._C._accelerator_setAllocatorSettings (#…
guangyey Oct 17, 2025
1ba808d
Refine CUDA BackendStaticInitializer for allocator select (#165298)
guangyey Oct 17, 2025
1b121d6
Fix AllocatorConfig parse roundup division bug (#165304)
guangyey Oct 17, 2025
22ae059
AOTI util deprecated flow using the new tracer (#165582)
tugsbayasgalan Oct 17, 2025
c73f508
Migrating some more callsites (#163580)
tugsbayasgalan Oct 17, 2025
15ff1cd
Remove E721 suppression in flake8 (#165855)
cyyever Oct 19, 2025
fa0db21
shrink_group implementation to expose ncclCommShrink API (#164518)
brchang24 Oct 19, 2025
633a3b7
Revert "shrink_group implementation to expose ncclCommShrink API (#16…
pytorchmergebot Oct 19, 2025
a885873
[dynamo] Clean up assert in dynamo [1/N] (#165430)
can-gaa-hou Oct 19, 2025
8139f33
[dynamo] Add recompile reason for set_stance fail_on_recompile (#165445)
parsshar-RH Oct 19, 2025
8951df0
test_scaled_matmul_cuda: fix infer_scale_swizzle (#165788)
jagadish-amd Oct 19, 2025
6b80c94
[FlexAttention] Fix dynamic shaped heads flex_flash check (#165866)
drisspg Oct 19, 2025
8a8329b
[ATen] Switch order of blocked reduce when vectorize loads (#165178)
PaulZhang12 Oct 17, 2025
61d9a51
[Fix XPU CI] [Inductor UT] Fix test cases broken by community. (#165…
etaf Oct 19, 2025
e9d8973
[Inductor] support masked vectorization for the tail_loop for float64…
jiayisunx Oct 15, 2025
e8cb34d
[Inductor] support masked vectorization for the tail_loop for fp8 dat…
jiayisunx Oct 15, 2025
47804ce
Revert "12/n : Remove fbandroid_compiler_flags (#165558)"
pytorchmergebot Oct 20, 2025
602ace5
Revert "[ATen] Fix CUDA reduction warp shuffle order (#164790)"
pytorchmergebot Oct 20, 2025
767199f
[flex_attention] replace sliced BlockMask noop with helpful error (#1…
Aminsed Oct 20, 2025
e1e8491
[1/N] Change C-style casts to static_cast or reinterpret_cast (#165750)
cyyever Oct 20, 2025
722b2b8
[dynamo] Remove duplicated guards (#165806)
anijain2305 Oct 18, 2025
ba93d56
[cuda] fix nll_loss2d backward bounds check with reduction=none (#165…
Aminsed Oct 20, 2025
c1eda34
[cuda] fix triu/tril int32 overflow for large matrices (#164705)
Aminsed Oct 20, 2025
2705937
[CI] Add rocm CI back to trunk for pre-submit/PR jobs (#165674)
jithunnair-amd Oct 20, 2025
b23f468
[Inductor][CuTeDSL] Move load_template up two directories (#165868)
NikhilAPatel Oct 20, 2025
ab82456
Revert "[1/N] Change C-style casts to static_cast or reinterpret_cast…
pytorchmergebot Oct 20, 2025
8700d68
[ROCm][CI] Update rocm.yml workflow to use 1 GPU ARC runners (#165481)
amdfaa Oct 20, 2025
0da1f91
Revert "[Submodule] Bump FBGEMM to latest (#165544)"
pytorchmergebot Oct 20, 2025
62a263b
Revert "Widen ops support to take in IntHOArrayRef vs only std::vec (…
pytorchmergebot Oct 20, 2025
69a4bfe
Revert "Refactor out headeronly ArrayRef (#164991)"
pytorchmergebot Oct 20, 2025
9875e70
Revert "[dynamo][misc] Replace UserFunctionVariable with VariableTrac…
pytorchmergebot Oct 20, 2025
0bf6043
Revert "[dynamo][user_defined] Replace UserFunctionVariable with Vari…
pytorchmergebot Oct 20, 2025
ca7360e
Revert "Move toString(ScalarType) and ScalarType ostream operator to …
pytorchmergebot Oct 20, 2025
150682b
Revert "Remove workaround to old CUDA bug (#164354)"
pytorchmergebot Oct 20, 2025
240c133
Revert "[inductor] require shape in TritonCSEVariable (#162275)"
pytorchmergebot Oct 20, 2025
8f06a13
[MPS] slightly faster cholesky (#165867)
Isalia20 Oct 20, 2025
779296a
[Inductor] Naive foreach autotune support (#162053)
jataylo Oct 20, 2025
4f7f432
Revert "[ROCm][CI] Update rocm.yml workflow to use 1 GPU ARC runners …
pytorchmergebot Oct 20, 2025
efc277c
[annotation] add logging for debugging annotation (#165797)
yushangdi Oct 20, 2025
cf280ca
Revert "[Inductor] Naive foreach autotune support (#162053)"
pytorchmergebot Oct 20, 2025
1891239
[Graph Partition] fix graph partition input signature for fallback ke…
BoyuanFeng Oct 20, 2025
df64c0c
[Code Clean] Clean asserts in torch/ao/quantization (root, quantizer,…
zhudada0120 Oct 20, 2025
1bcd736
fix bad merge duplicate pre pass (#165917)
eellison Oct 20, 2025
850ba8c
[Code Clean] Clean asserts in torch/autograd. (#165627)
mansiag05 Oct 20, 2025
96b0e7a
[Code Clean] Clean asserts in torch/ao/quantization/experimental/* an…
zhudada0120 Oct 20, 2025
99c8640
[1/N] Change C-style casts to static_cast or reinterpret_cast (#165750)
cyyever Oct 20, 2025
e20c9bf
[torch/utils][Code Clean] Clean asserts in `torch/utils/*.py` (#165410)
KarhouTam Oct 20, 2025
259cb94
[stage 2c] make autograd and inference functions (#165668)
avikchaudhuri Oct 20, 2025
70592c6
[ROCm][CI] Move gfx1100 workflows to own yaml file (#165699)
jithunnair-amd Oct 20, 2025
51d0d8e
[ATen] Fix CUDA reduction warp shuffle order (#164790)
PaulZhang12 Oct 20, 2025
b20deec
[PP] Add optional argument to not save outputs (#165822)
H-Huang Oct 18, 2025
4c963a6
Use inline instead of anon namespace for stableivalue from/to (#164882)
janeyx99 Oct 20, 2025
4a6cf0a
Fix dynamo stack trace (#165930)
yushangdi Oct 21, 2025
0b1c462
Making Numpy depedency in Local Tensor optional to fix broken Torchao…
dzmitry-huba Oct 20, 2025
d7d4bb7
Add XPU part for persons_of_interest (#165920)
EikanWang Oct 20, 2025
303c9cf
Save Python refcount bump on each arg in maybe_handle_torch_function …
swolchok Oct 20, 2025
7406d2e
[DeviceMesh] Clean up the call into mesh_resouces to get root mesh (#…
fduwjj Oct 20, 2025
0be0de4
Add type suppressions to _inductor/runtime (#165918)
maggiemoss Oct 21, 2025
fe69a2b
Move from/to to torch::stable::detail (#164956)
janeyx99 Oct 21, 2025
a238a9a
Add clang-tidy misc-definitions-in-headers check (#164959)
janeyx99 Oct 21, 2025
ce1fcff
[ROCm] Keep amdgpu-coerce-illegal-types flag if rocm version is less …
ethanwee1 Oct 21, 2025
0e08394
Enable PLW0127 in ruff (#165851)
cyyever Oct 21, 2025
ffa90d4
[ROCm][CI] Update rocm.yml workflow to use 1 GPU ARC runners (#165481)
amdfaa Oct 21, 2025
7711708
[dynamo][nn_module] Replace UserFunctionVariable with VariableTracker…
anijain2305 Oct 21, 2025
03f3f78
[ATen] Add reduction tag to reduction operators (#165155)
eellison Oct 21, 2025
f2bb22f
[Inductor-FX] Support Tensor.item (#165599)
blaine-rister Oct 21, 2025
485cef3
support high stream
Chao1Han Sep 15, 2025
0891988
fix
Chao1Han Sep 15, 2025
65d78d2
update
Chao1Han Sep 15, 2025
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
10 changes: 6 additions & 4 deletions .ci/aarch64_linux/aarch64_ci_build.sh
Original file line number Diff line number Diff line change
Expand Up @@ -8,13 +8,17 @@ if [[ "$GPU_ARCH_VERSION" == *"12.6"* ]]; then
export TORCH_CUDA_ARCH_LIST="8.0;9.0"
elif [[ "$GPU_ARCH_VERSION" == *"12.8"* ]]; then
export TORCH_CUDA_ARCH_LIST="8.0;9.0;10.0;12.0"
elif [[ "$GPU_ARCH_VERSION" == *"12.9"* ]]; then
export TORCH_CUDA_ARCH_LIST="8.0;9.0;10.0;12.0"
elif [[ "$GPU_ARCH_VERSION" == *"13.0"* ]]; then
export TORCH_CUDA_ARCH_LIST="8.0;9.0;10.0;11.0;12.0+PTX"
fi

# Compress the fatbin with -compress-mode=size for CUDA 13
if [[ "$DESIRED_CUDA" == *"13"* ]]; then
export TORCH_NVCC_FLAGS="-compress-mode=size"
# Bundle ptxas into the cu13 wheel, see https://github.com/pytorch/pytorch/issues/163801
export BUILD_BUNDLE_PTXAS=1
fi

SCRIPTPATH="$( cd -- "$(dirname "$0")" >/dev/null 2>&1 ; pwd -P )"
Expand All @@ -31,8 +35,7 @@ pip install -r /pytorch/requirements.txt
pip install auditwheel==6.2.0 wheel
if [ "$DESIRED_CUDA" = "cpu" ]; then
echo "BASE_CUDA_VERSION is not set. Building cpu wheel."
#USE_PRIORITIZED_TEXT_FOR_LD for enable linker script optimization https://github.com/pytorch/pytorch/pull/121975/files
USE_PRIORITIZED_TEXT_FOR_LD=1 python /pytorch/.ci/aarch64_linux/aarch64_wheel_ci_build.py --enable-mkldnn
python /pytorch/.ci/aarch64_linux/aarch64_wheel_ci_build.py --enable-mkldnn
else
echo "BASE_CUDA_VERSION is set to: $DESIRED_CUDA"
export USE_SYSTEM_NCCL=1
Expand All @@ -46,6 +49,5 @@ else
export USE_NVIDIA_PYPI_LIBS=1
fi

#USE_PRIORITIZED_TEXT_FOR_LD for enable linker script optimization https://github.com/pytorch/pytorch/pull/121975/files
USE_PRIORITIZED_TEXT_FOR_LD=1 python /pytorch/.ci/aarch64_linux/aarch64_wheel_ci_build.py --enable-mkldnn --enable-cuda
python /pytorch/.ci/aarch64_linux/aarch64_wheel_ci_build.py --enable-mkldnn --enable-cuda
fi
59 changes: 5 additions & 54 deletions .ci/aarch64_linux/aarch64_wheel_ci_build.py
Original file line number Diff line number Diff line change
Expand Up @@ -13,49 +13,6 @@ def list_dir(path: str) -> list[str]:
return check_output(["ls", "-1", path]).decode().split("\n")


def build_ArmComputeLibrary() -> None:
"""
Using ArmComputeLibrary for aarch64 PyTorch
"""
print("Building Arm Compute Library")
acl_build_flags = [
"debug=0",
"neon=1",
"opencl=0",
"os=linux",
"openmp=1",
"cppthreads=0",
"arch=armv8a",
"multi_isa=1",
"fixed_format_kernels=1",
"build=native",
]
acl_install_dir = "/acl"
acl_checkout_dir = os.getenv("ACL_SOURCE_DIR", "ComputeLibrary")
if os.path.isdir(acl_install_dir):
shutil.rmtree(acl_install_dir)
if not os.path.isdir(acl_checkout_dir) or not len(os.listdir(acl_checkout_dir)):
check_call(
[
"git",
"clone",
"https://github.com/ARM-software/ComputeLibrary.git",
"-b",
"v25.02",
"--depth",
"1",
"--shallow-submodules",
]
)

check_call(
["scons", "Werror=1", f"-j{os.cpu_count()}"] + acl_build_flags,
cwd=acl_checkout_dir,
)
for d in ["arm_compute", "include", "utils", "support", "src", "build"]:
shutil.copytree(f"{acl_checkout_dir}/{d}", f"{acl_install_dir}/{d}")


def replace_tag(filename) -> None:
with open(filename) as f:
lines = f.readlines()
Expand Down Expand Up @@ -317,7 +274,7 @@ def parse_arguments():
).decode()

print("Building PyTorch wheel")
build_vars = "CMAKE_SHARED_LINKER_FLAGS=-Wl,-z,max-page-size=0x10000 "
build_vars = ""
# MAX_JOB=5 is not required for CPU backend (see commit 465d98b)
if enable_cuda:
build_vars += "MAX_JOBS=5 "
Expand Down Expand Up @@ -356,23 +313,17 @@ def parse_arguments():
build_vars += f"BUILD_TEST=0 PYTORCH_BUILD_VERSION={branch[1 : branch.find('-')]} PYTORCH_BUILD_NUMBER=1 "

if enable_mkldnn:
build_ArmComputeLibrary()
print("build pytorch with mkldnn+acl backend")
build_vars += (
"USE_MKLDNN=ON USE_MKLDNN_ACL=ON "
"ACL_ROOT_DIR=/acl "
"LD_LIBRARY_PATH=/pytorch/build/lib:/acl/build:$LD_LIBRARY_PATH "
"ACL_INCLUDE_DIR=/acl/build "
"ACL_LIBRARY=/acl/build "
)
build_vars += "USE_MKLDNN=ON USE_MKLDNN_ACL=ON "
build_vars += "ACL_ROOT_DIR=/acl "
if enable_cuda:
build_vars += "BLAS=NVPL "
else:
build_vars += "BLAS=OpenBLAS OpenBLAS_HOME=/OpenBLAS "
build_vars += "BLAS=OpenBLAS OpenBLAS_HOME=/opt/OpenBLAS "
else:
print("build pytorch without mkldnn backend")

os.system(f"cd /pytorch; {build_vars} python3 setup.py bdist_wheel")
os.system(f"cd /pytorch; {build_vars} python3 -m build --wheel --no-isolation")
if enable_cuda:
print("Updating Cuda Dependency")
filename = os.listdir("/pytorch/dist/")
Expand Down
64 changes: 17 additions & 47 deletions .ci/aarch64_linux/build_aarch64_wheel.py
Original file line number Diff line number Diff line change
Expand Up @@ -241,7 +241,7 @@ def wait_for_connection(addr, port, timeout=15, attempt_cnt=5):
try:
with socket.create_connection((addr, port), timeout=timeout):
return
except (ConnectionRefusedError, socket.timeout): # noqa: PERF203
except (ConnectionRefusedError, TimeoutError): # noqa: PERF203
if i == attempt_cnt - 1:
raise
time.sleep(timeout)
Expand Down Expand Up @@ -299,40 +299,6 @@ def install_condaforge_python(host: RemoteHost, python_version="3.8") -> None:
)


def build_OpenBLAS(host: RemoteHost, git_clone_flags: str = "") -> None:
print("Building OpenBLAS")
host.run_cmd(
f"git clone https://github.com/xianyi/OpenBLAS -b v0.3.28 {git_clone_flags}"
)
make_flags = "NUM_THREADS=64 USE_OPENMP=1 NO_SHARED=1 DYNAMIC_ARCH=1 TARGET=ARMV8"
host.run_cmd(
f"pushd OpenBLAS && make {make_flags} -j8 && sudo make {make_flags} install && popd && rm -rf OpenBLAS"
)


def build_ArmComputeLibrary(host: RemoteHost, git_clone_flags: str = "") -> None:
print("Building Arm Compute Library")
acl_build_flags = " ".join(
[
"debug=0",
"neon=1",
"opencl=0",
"os=linux",
"openmp=1",
"cppthreads=0",
"arch=armv8a",
"multi_isa=1",
"fixed_format_kernels=1",
"build=native",
]
)
host.run_cmd(
f"git clone https://github.com/ARM-software/ComputeLibrary.git -b v25.02 {git_clone_flags}"
)

host.run_cmd(f"cd ComputeLibrary && scons Werror=1 -j8 {acl_build_flags}")


def embed_libgomp(host: RemoteHost, use_conda, wheel_name) -> None:
host.run_cmd("pip3 install auditwheel")
host.run_cmd(
Expand Down Expand Up @@ -442,7 +408,7 @@ def build_torchvision(
if host.using_docker():
build_vars += " CMAKE_SHARED_LINKER_FLAGS=-Wl,-z,max-page-size=0x10000"

host.run_cmd(f"cd vision && {build_vars} python3 setup.py bdist_wheel")
host.run_cmd(f"cd vision && {build_vars} python3 -m build --wheel --no-isolation")
vision_wheel_name = host.list_dir("vision/dist")[0]
embed_libgomp(host, use_conda, os.path.join("vision", "dist", vision_wheel_name))

Expand Down Expand Up @@ -497,7 +463,7 @@ def build_torchdata(
if host.using_docker():
build_vars += " CMAKE_SHARED_LINKER_FLAGS=-Wl,-z,max-page-size=0x10000"

host.run_cmd(f"cd data && {build_vars} python3 setup.py bdist_wheel")
host.run_cmd(f"cd data && {build_vars} python3 -m build --wheel --no-isolation")
wheel_name = host.list_dir("data/dist")[0]
embed_libgomp(host, use_conda, os.path.join("data", "dist", wheel_name))

Expand Down Expand Up @@ -553,7 +519,7 @@ def build_torchtext(
if host.using_docker():
build_vars += " CMAKE_SHARED_LINKER_FLAGS=-Wl,-z,max-page-size=0x10000"

host.run_cmd(f"cd text && {build_vars} python3 setup.py bdist_wheel")
host.run_cmd(f"cd text && {build_vars} python3 -m build --wheel --no-isolation")
wheel_name = host.list_dir("text/dist")[0]
embed_libgomp(host, use_conda, os.path.join("text", "dist", wheel_name))

Expand Down Expand Up @@ -614,7 +580,7 @@ def build_torchaudio(
host.run_cmd(
f"cd audio && export FFMPEG_ROOT=$(pwd)/third_party/ffmpeg && export USE_FFMPEG=1 \
&& ./packaging/ffmpeg/build.sh \
&& {build_vars} python3 setup.py bdist_wheel"
&& {build_vars} python3 -m build --wheel --no-isolation"
)

wheel_name = host.list_dir("audio/dist")[0]
Expand Down Expand Up @@ -700,7 +666,6 @@ def start_build(
configure_system(
host, compiler=compiler, use_conda=use_conda, python_version=python_version
)
build_OpenBLAS(host, git_clone_flags)

if host.using_docker():
print("Move libgfortant.a into a standard location")
Expand All @@ -723,10 +688,12 @@ def start_build(
f"git clone --recurse-submodules -b {branch} https://github.com/pytorch/pytorch {git_clone_flags}"
)

host.run_cmd("pytorch/.ci/docker/common/install_openblas.sh")

print("Building PyTorch wheel")
build_opts = ""
if pytorch_build_number is not None:
build_opts += f" --build-number {pytorch_build_number}"
build_opts += f" -C--build-option=--build-number={pytorch_build_number}"
# Breakpad build fails on aarch64
build_vars = "USE_BREAKPAD=0 "
if branch == "nightly":
Expand All @@ -743,15 +710,18 @@ def start_build(
if host.using_docker():
build_vars += " CMAKE_SHARED_LINKER_FLAGS=-Wl,-z,max-page-size=0x10000"
if enable_mkldnn:
build_ArmComputeLibrary(host, git_clone_flags)
host.run_cmd("pytorch/.ci/docker/common/install_acl.sh")
print("build pytorch with mkldnn+acl backend")
build_vars += " USE_MKLDNN=ON USE_MKLDNN_ACL=ON"
build_vars += " BLAS=OpenBLAS"
build_vars += " OpenBLAS_HOME=/opt/OpenBLAS"
build_vars += " ACL_ROOT_DIR=/acl"
host.run_cmd(
f"cd $HOME/pytorch && export ACL_ROOT_DIR=$HOME/ComputeLibrary && {build_vars} python3 setup.py bdist_wheel{build_opts}"
f"cd $HOME/pytorch && {build_vars} python3 -m build --wheel --no-isolation{build_opts}"
)
print("Repair the wheel")
pytorch_wheel_name = host.list_dir("pytorch/dist")[0]
ld_library_path = "$HOME/acl/build:$HOME/pytorch/build/lib"
ld_library_path = "/acl/build:$HOME/pytorch/build/lib"
host.run_cmd(
f"export LD_LIBRARY_PATH={ld_library_path} && auditwheel repair $HOME/pytorch/dist/{pytorch_wheel_name}"
)
Expand All @@ -763,7 +733,7 @@ def start_build(
else:
print("build pytorch without mkldnn backend")
host.run_cmd(
f"cd pytorch && {build_vars} python3 setup.py bdist_wheel{build_opts}"
f"cd pytorch && {build_vars} python3 -m build --wheel --no-isolation{build_opts}"
)

print("Deleting build folder")
Expand Down Expand Up @@ -907,7 +877,7 @@ def terminate_instances(instance_type: str) -> None:
def parse_arguments():
from argparse import ArgumentParser

parser = ArgumentParser("Builid and test AARCH64 wheels using EC2")
parser = ArgumentParser("Build and test AARCH64 wheels using EC2")
parser.add_argument("--key-name", type=str)
parser.add_argument("--debug", action="store_true")
parser.add_argument("--build-only", action="store_true")
Expand Down Expand Up @@ -1004,7 +974,7 @@ def parse_arguments():
install_condaforge_python(host, args.python_version)
sys.exit(0)

python_version = args.python_version if args.python_version is not None else "3.9"
python_version = args.python_version if args.python_version is not None else "3.10"

if args.use_torch_from_pypi:
configure_system(host, compiler=args.compiler, python_version=python_version)
Expand Down
3 changes: 2 additions & 1 deletion .ci/docker/almalinux/Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -69,7 +69,8 @@ RUN bash ./install_cuda.sh 13.0
ENV DESIRED_CUDA=13.0

FROM ${ROCM_IMAGE} as rocm
ENV PYTORCH_ROCM_ARCH="gfx900;gfx906;gfx908;gfx90a;gfx942;gfx1030;gfx1100;gfx1101;gfx1102;gfx1200;gfx1201"
ARG PYTORCH_ROCM_ARCH
ENV PYTORCH_ROCM_ARCH ${PYTORCH_ROCM_ARCH}
ADD ./common/install_mkl.sh install_mkl.sh
RUN bash ./install_mkl.sh && rm install_mkl.sh
ENV MKLROOT /opt/intel
Expand Down
6 changes: 6 additions & 0 deletions .ci/docker/almalinux/build.sh
Original file line number Diff line number Diff line change
Expand Up @@ -36,6 +36,12 @@ case ${DOCKER_TAG_PREFIX} in
;;
rocm*)
BASE_TARGET=rocm
PYTORCH_ROCM_ARCH="gfx900;gfx906;gfx908;gfx90a;gfx942;gfx1030;gfx1100;gfx1101;gfx1102;gfx1200;gfx1201"
# add gfx950, gfx115x conditionally starting in ROCm 7.0
if [[ "$ROCM_VERSION" == *"7.0"* ]]; then
PYTORCH_ROCM_ARCH="${PYTORCH_ROCM_ARCH};gfx950;gfx1150;gfx1151"
fi
EXTRA_BUILD_ARGS="${EXTRA_BUILD_ARGS} --build-arg PYTORCH_ROCM_ARCH=${PYTORCH_ROCM_ARCH}"
;;
*)
echo "ERROR: Unknown docker tag ${DOCKER_TAG_PREFIX}"
Expand Down
41 changes: 10 additions & 31 deletions .ci/docker/build.sh
Original file line number Diff line number Diff line change
Expand Up @@ -84,8 +84,8 @@ fi
_UCX_COMMIT=7836b165abdbe468a2f607e7254011c07d788152
_UCC_COMMIT=430e241bf5d38cbc73fc7a6b89155397232e3f96
if [[ "$image" == *rocm* ]]; then
_UCX_COMMIT=cc312eaa4655c0cc5c2bcd796db938f90563bcf6
_UCC_COMMIT=0c0fc21559835044ab107199e334f7157d6a0d3d
_UCX_COMMIT=29831d319e6be55cb8c768ca61de335c934ca39e
_UCC_COMMIT=9f4b242cbbd8b1462cbc732eb29316cdfa124b77
fi

tag=$(echo $image | awk -F':' '{print $2}')
Expand Down Expand Up @@ -113,6 +113,7 @@ case "$tag" in
UCX_COMMIT=${_UCX_COMMIT}
UCC_COMMIT=${_UCC_COMMIT}
TRITON=yes
INSTALL_MINGW=yes
;;
pytorch-linux-jammy-cuda13.0-cudnn9-py3-gcc11)
CUDA_VERSION=13.0.0
Expand Down Expand Up @@ -175,28 +176,17 @@ case "$tag" in
fi
GCC_VERSION=11
VISION=yes
ROCM_VERSION=6.4
ROCM_VERSION=7.0
NINJA_VERSION=1.9.0
TRITON=yes
KATEX=yes
UCX_COMMIT=${_UCX_COMMIT}
UCC_COMMIT=${_UCC_COMMIT}
PYTORCH_ROCM_ARCH="gfx90a;gfx942;gfx950;gfx1100"
if [[ $tag =~ "benchmarks" ]]; then
INDUCTOR_BENCHMARKS=yes
fi
;;
pytorch-linux-noble-rocm-alpha-py3)
ANACONDA_PYTHON_VERSION=3.12
GCC_VERSION=11
VISION=yes
ROCM_VERSION=7.0
NINJA_VERSION=1.9.0
TRITON=yes
KATEX=yes
UCX_COMMIT=${_UCX_COMMIT}
UCC_COMMIT=${_UCC_COMMIT}
PYTORCH_ROCM_ARCH="gfx90a;gfx942;gfx950"
;;
pytorch-linux-jammy-xpu-n-1-py3)
ANACONDA_PYTHON_VERSION=3.10
GCC_VERSION=11
Expand Down Expand Up @@ -262,13 +252,10 @@ case "$tag" in
TRITON_CPU=yes
;;
pytorch-linux-jammy-linter)
# TODO: Use 3.9 here because of this issue https://github.com/python/mypy/issues/13627.
# We will need to update mypy version eventually, but that's for another day. The task
# would be to upgrade mypy to 1.0.0 with Python 3.11
PYTHON_VERSION=3.9
PYTHON_VERSION=3.10
;;
pytorch-linux-jammy-cuda12.8-cudnn9-py3.9-linter)
PYTHON_VERSION=3.9
pytorch-linux-jammy-cuda12.8-cudnn9-py3.10-linter)
PYTHON_VERSION=3.10
CUDA_VERSION=12.8.1
;;
pytorch-linux-jammy-aarch64-py3.10-gcc11)
Expand Down Expand Up @@ -358,7 +345,7 @@ docker build \
--build-arg "NINJA_VERSION=${NINJA_VERSION:-}" \
--build-arg "KATEX=${KATEX:-}" \
--build-arg "ROCM_VERSION=${ROCM_VERSION:-}" \
--build-arg "PYTORCH_ROCM_ARCH=${PYTORCH_ROCM_ARCH:-gfx90a;gfx942}" \
--build-arg "PYTORCH_ROCM_ARCH=${PYTORCH_ROCM_ARCH}" \
--build-arg "IMAGE_NAME=${IMAGE_NAME}" \
--build-arg "UCX_COMMIT=${UCX_COMMIT}" \
--build-arg "UCC_COMMIT=${UCC_COMMIT}" \
Expand All @@ -375,6 +362,7 @@ docker build \
--build-arg "OPENBLAS=${OPENBLAS:-}" \
--build-arg "SKIP_SCCACHE_INSTALL=${SKIP_SCCACHE_INSTALL:-}" \
--build-arg "SKIP_LLVM_SRC_BUILD_INSTALL=${SKIP_LLVM_SRC_BUILD_INSTALL:-}" \
--build-arg "INSTALL_MINGW=${INSTALL_MINGW:-}" \
-f $(dirname ${DOCKERFILE})/Dockerfile \
-t "$tmp_tag" \
"$@" \
Expand Down Expand Up @@ -455,12 +443,3 @@ elif [ "$HAS_TRITON" = "yes" ]; then
echo "expecting triton to not be installed, but it is"
exit 1
fi

# Sanity check cmake version. Executorch reinstalls cmake and I'm not sure if
# they support 4.0.0 yet, so exclude them from this check.
CMAKE_VERSION=$(drun cmake --version)
if [[ "$EXECUTORCH" != *yes* && "$CMAKE_VERSION" != *4.* ]]; then
echo "CMake version is not 4.0.0:"
drun cmake --version
exit 1
fi
Loading