Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
468 commits
Select commit Hold shift + click to select a range
79eca46
[precompile] Skip serializing unnecesssary objects for guards. (#158926)
zhxchen17 Aug 6, 2025
d87161c
[Easy] Fix wrong propagation of fallback_ops_dict in gen_aoti_c_shim …
mikaylagawarecki Aug 5, 2025
a4b07fe
[AOTI] Add more default options to compile_standalone (#158560)
desertfire Aug 6, 2025
4c01991
[DCP][Prototype] Checkpoint replication via PGTransport (#157963) (#1…
meetv18 Aug 6, 2025
d7a855d
[async-TP] Make scaled-mm + reduce-scatter preserve alignment of scal…
lw Aug 6, 2025
c669b0a
Fix execution frame cleanup logic (#158717)
georgiaphillips Aug 6, 2025
44dd368
[AOTI] Fix memory leak from all_reduce (#159818)
desertfire Aug 4, 2025
ba37f58
Revert "[dynamo] Be consistent with storing func source for UserMetho…
pytorchmergebot Aug 6, 2025
6fa3592
Dataloader benchmark script (#159432)
divyanshk Aug 6, 2025
c719509
[inductor] add _get_inductor_debug_symbol_cflags for debug symbol con…
xuhancn Aug 6, 2025
d10e9e4
[MPS] Remove all pre-MacOS14 logic (#159912)
malfet Aug 6, 2025
12a54e4
[Inductor UT][Fix XPU CI] Fix case failures introduced by community. …
etaf Aug 6, 2025
0de2a45
[BE] Merge 3 CUDA build jobs into one (#159890)
malfet Aug 6, 2025
b8ef60b
Enable XNNPACK aarch64 builds (#159762)
frankseide Aug 6, 2025
50580b5
Add minimal nn.functional.log_softmax support for NestedTensor (#159662)
alanhdu Aug 6, 2025
0afaeb7
Improve `extract_test_fn` (#158637)
kundaMwiza Aug 6, 2025
d2368aa
[CPUBLAS] add macros for brgemm APIs for versioning (#158629)
Xia-Weiwen Aug 6, 2025
512b473
[EZ] Remove useless `cross_compile_arm64` (#159986)
malfet Aug 6, 2025
289f62c
[inductor][ez] fixup scaled_mm (#159948)
coconutruben Aug 6, 2025
a572596
Remove unnecessary "# noqa: set_linter" comments (#159467)
rec Aug 6, 2025
40c4d61
[Dynamo][Better Engineering] Typing `torch/_dynamo/guards.py` (#159315)
Lucaskabela Aug 6, 2025
2507ae6
Partitioner: Fix to align partition node order with original graph (#…
xwu-intel Aug 6, 2025
9fd5b5f
[pytorch] Moving torch.compile worker process logs to a dedicated ran…
jovianjaison Aug 6, 2025
3a2c3c8
unskipped mobilenet_v3 quantization and mobilenet_v2 quantization plu…
christinaburge Aug 6, 2025
93da995
gloo: fix building system gloo with CUDA/HIP (#146637)
nlbrown2 Aug 6, 2025
64dc30c
[HOP, map] Rework of map autograd to the new interface (#153343)
bohnstingl Aug 6, 2025
a6bc296
[FlexAttention] Update the guard semantics for divisibility (#159884)
drisspg Aug 5, 2025
cb4b29b
Revert "[pytorch] Moving torch.compile worker process logs to a dedic…
pytorchmergebot Aug 6, 2025
3daef4d
[dynamo] Trace nn.Module __delattr__ (#159969)
anijain2305 Aug 6, 2025
fd606a3
[dynamo] update pytorch-labs -> meta-pytorch in graph break URLs (#15…
williamwen42 Aug 6, 2025
5cedc5a
[BE][PYFMT] migrate PYFMT for `torch/[p-z]*/` to `ruff format` (#144552)
XuehaiPan Aug 6, 2025
8b0be7b
[Profiler] Fix unexpected C return events (#159574)
D-D-H Aug 7, 2025
1bb5e6c
update expected results (#159867)
laithsakka Aug 5, 2025
2ba2f59
[Dynamo] Add torch.xpu.stream to trace rules (#159844)
guangyey Aug 5, 2025
38d65c6
Add a USE_NIGHTLY option to setup.py (#159965)
ezyang Aug 6, 2025
d022671
[BE][EZ] Delete remains of split-build logic (#159990)
malfet Aug 6, 2025
81d72fb
Move smoke binary builds to 3.12 (#159993)
malfet Aug 6, 2025
d4c1a08
Relax unclaimed successes in dtype op tests when running under TEST_W…
xmfan Aug 6, 2025
c859ba7
Make onnx export SDPA match aten behavior (#159973)
IlyasMoutawwakil Aug 7, 2025
3f1636e
[audio hash update] update the pinned audio hash (#160046)
pytorchupdatebot Aug 7, 2025
aa75e91
[Export Schema] Remove deviceAllocationMap field (#159653)
SherlockNoMad Aug 7, 2025
24f43d0
[inductor] [cpu] fix the dype hardcoded to int64 in store_reduction (…
thenumberouscode Aug 7, 2025
422bd68
dataclass pytree fix (#159916)
avikchaudhuri Aug 7, 2025
b0df771
Remove benchmark dependencies from regular ROCm CI images (#160047)
huydhn Aug 7, 2025
8cb91e2
Renaming HAS_XPU to HAS_XPU_AND_TRITON (#159908)
dsashidh Aug 7, 2025
a53d14d
Revert "unskipped mobilenet_v3 quantization and mobilenet_v2 quantiza…
pytorchmergebot Aug 7, 2025
83875cd
[nativert] Expose ModelRunner to public through pmpl type ModelRunner…
zhxchen17 Aug 7, 2025
d20c4c2
[CI] Update xpu ci use rolling driver for new features (#158340)
chuanqi129 Aug 7, 2025
8ab5868
Actually run the einops tests in CI (#159776)
zou3519 Aug 4, 2025
f60454c
S390X: update test dependencies (#158636)
AlekseiNikiforovIBM Aug 7, 2025
e248719
[DTensor] support _StridedShard in view op (#159656)
XilunWu Aug 7, 2025
90b78ee
Move xla jobs to unstable workflow (#159272)
ZainRizvi Aug 7, 2025
c4e6446
Revert "Add UT for torch.accelerator memory-related API (#155200)"
pytorchmergebot Aug 7, 2025
74da260
Revert "Add unified memory APIs for torch.accelerator (#152932)"
pytorchmergebot Aug 7, 2025
f3a4d74
Revert "Add DeviceAllocator as the base device allocator (#138222)"
pytorchmergebot Aug 7, 2025
06824f3
[inductor] fix test_dynamo_timed on Windows. (#159981)
xuhancn Aug 7, 2025
e1cf0d4
[inductor] unification for inductor debug. (#159998)
xuhancn Aug 7, 2025
b1a6027
[Profiler] Update README (#159816)
sraikund16 Aug 7, 2025
e167c7d
[inductor] allocate non-blocking copy destinations in pinned memory (…
v0i0 Aug 6, 2025
57f738b
[inductor] move all cpu scalars using pinned memory for graph partiti…
v0i0 Aug 6, 2025
69cc606
HF component update to not use fsspec components (#159405)
ankitageorge Aug 7, 2025
0b187b3
DCP HF reader: use safe_open instead of reading the bytes (#159406)
ankitageorge Aug 7, 2025
8399cf8
Use only safetensors APIs in HFStorageReader (#159681)
ankitageorge Aug 7, 2025
0bd3af4
Further fix failing tests in test/inductor/test_analysis.py (#160070)
Aidyn-A Aug 7, 2025
ee1fb43
Fix docker image creation (#158634)
AlekseiNikiforovIBM Aug 7, 2025
21392c0
[inductor] disable flex decoding on Windows. (#160072)
xuhancn Aug 7, 2025
3cf7b40
[DTensor] Support user-supplied Generator for random ops (#159933)
wconstab Aug 6, 2025
e619c6b
[export] Apply move_to_device_pass to all submodules (#159992)
angelayi Aug 7, 2025
8147370
Fix qembeddingbag_byte_prepack_meta to use sym_sizes (#159985)
SherlockNoMad Aug 7, 2025
36f46d0
dynamo: Remove passing or deleted dynamo_expected_failures (#159691)
c00w Aug 5, 2025
d46768d
[MTIA] Allow users who know what they are doing to ignore all device …
patrick-toulme Aug 7, 2025
f077c24
[replicate][be] improved readability of test case description (#160128)
anshul-si Aug 7, 2025
195b5c2
Revert "dynamo: Remove passing or deleted dynamo_expected_failures (#…
pytorchmergebot Aug 7, 2025
03b254e
Extend torch function support to ALL arguments, not just scalar type …
ezyang Aug 7, 2025
d68c323
Log max_autotune exceptions (#159687) (#159688)
wychi Aug 8, 2025
ba4ccf5
turn on executon frame clenaup by default (#160110)
georgiaphillips Aug 8, 2025
05c4177
integrate kernacle into inductor (#160121)
bobrenjc93 Aug 7, 2025
3fcd79e
Fix infinite loop when iterating over an empty zip (#159673)
guilhermeleobas Aug 1, 2025
beb4d78
[BE]: ruff PLC0207 - use maxsplit kwarg (#160107)
Skylion007 Aug 8, 2025
2ea40fb
[Linter] Improve device-bias linter by adding detection for `with tor…
etaf Aug 6, 2025
017259f
[benchmarks] Add nativert benchmark (#159922)
yiming0416 Aug 8, 2025
24257f5
[vllm hash update] update the pinned vllm hash (#159822)
pytorchupdatebot Aug 8, 2025
b5c9372
[SymmMem] Add NVSHMEM Reduction support (sum, min, max) into Triton (…
codingwithsurya Aug 6, 2025
b0b229b
[SymmMem] Use _get_default_group() instead of group.WORLD for group_n…
codingwithsurya Aug 6, 2025
ea7fe0e
[SymmMem] Standardize NVSHMEM Triton wrappers on byte-based APIs + i…
codingwithsurya Aug 6, 2025
1783d6e
[SymmMem] Fix flaky wait_until test (#159215)
codingwithsurya Aug 6, 2025
7c4f7b9
[SymmMem] Add Triton 3.4 support to NVSHMEM Triton and fix CI tests …
codingwithsurya Aug 6, 2025
1c88144
[SymmMem] Initialize NVSHMEM module only for kernels that have nvshme…
codingwithsurya Aug 6, 2025
bfff2e3
[SymmMem] Refactor NVSHMEM Reduction API to be more ergonomic with au…
codingwithsurya Aug 8, 2025
e0d8a31
[SymmMem] Add helpful docstrings for all NVSHMEM APIs (#159756)
codingwithsurya Aug 8, 2025
3a56237
[SymmMem] Send tensors with unerased type information to NVSHMEM Trit…
codingwithsurya Aug 8, 2025
178515d
[BE][PYFMT] remove `black`: finish `black -> ruff format` migration (…
XuehaiPan Aug 7, 2025
556e2a7
[Test][Easy] Use float16 dtype in test_sort_large (#159939)
Aidyn-A Aug 8, 2025
7f4cb4a
[MPS] coalesce for sparse tensors (#159729)
Isalia20 Aug 8, 2025
62bac07
[inductor][triton] support profile_scratch launcher arg (#159772)
davidberard98 Aug 8, 2025
9fa8ce2
Working setup with runnable PyTorch on Codex. (#159968)
ezyang Aug 6, 2025
b5fd722
Improve pin_memory error message on CPU-only systems (#159994)
ezyang Aug 8, 2025
8a37f0c
improve gather and scatter_add strategy (#160140)
zpcore Aug 8, 2025
50f23ff
rename-HAS_CUDA-to-HAS_CUDA_AND_TRITON (#159883)
ghostspiders Aug 8, 2025
231c722
CMake build: preserve PYTHONPATH (#160144)
adamjstewart Aug 8, 2025
a4f69a5
[dynamo][guards] Remove guards on stdlib modules (#159913)
anijain2305 Aug 8, 2025
86eb65f
[MPS] Move max_pool2d to Metal for `stride != 1` (#157876)
kurtamohler Jul 31, 2025
c5ec545
Don't build nccl when distributed is disabled (#160086)
albanD Aug 8, 2025
d7114f0
Add DeviceAllocator as the base device allocator (#138222)
guangyey Aug 8, 2025
84f7e88
Add unified memory APIs for torch.accelerator (#152932)
guangyey Aug 8, 2025
da1f608
Add UT for torch.accelerator memory-related API (#155200)
guangyey Aug 8, 2025
5f5f508
[ROCm] Ck backend UX refactor (#152951)
alugorey Aug 8, 2025
72009ec
[replicate][be] improved readability and cleaned up remaining DDP cod…
anshul-si Aug 7, 2025
c86040a
[torch.export] Fix test_export_api_with_dynamic_shapes (#160164)
jd7-tr Aug 8, 2025
2ee22e4
[pytorch][dynamo_compile] Log stack_trace to dynamo_compile (#159655)
jovianjaison Aug 8, 2025
1febab2
Do not treat ReinterpretView as a realized node (#159920)
shengfukevin Aug 8, 2025
2247aa6
Documents tuning NVLink performance on H100/H200 (#159792)
syed-ahmed Aug 7, 2025
28ccc9e
[MPS] Extend `index_put` to complex types (#160159)
malfet Aug 8, 2025
206c1ee
Revert "[pytorch][dynamo_compile] Log stack_trace to dynamo_compile (…
pytorchmergebot Aug 8, 2025
334ecbd
Add torchao to install_inductor_benchmark_deps cleanup stage (#160191)
robert-hardwick Aug 8, 2025
1128f4c
[cuDNN][SDPA] cuDNN SDPA refactor/cleanup, nested tensor backward, te…
eqy Aug 8, 2025
9e07673
Fix test_fsdp_ep.py due to _MeshEnv API change (#158695)
Kanya-Mo Aug 8, 2025
4e2ddb5
[Inductor][CUTLASS] Copy cutlass_mock_imports directory (#159724)
Aidyn-A Aug 8, 2025
566c6d5
[ONNX] Fix the export of the model having none as output (#160200)
titaiwangms Aug 8, 2025
731ee31
[TorchScript, PT2] Add torch._check compatibility support (#159988)
gmagogsfm Aug 8, 2025
8c41cb8
[MPS][BE] Combine all pre-MacOS14 xfail lists (#160228)
malfet Aug 8, 2025
9b803cd
[BE] Remove more optim entries from docs coverage ignore list (#160194)
janeyx99 Aug 8, 2025
e96c7c4
[dcp][hf] Improve HF consolidation algorithm (#158648)
ankitageorge Aug 8, 2025
11a3565
[Torch Native] Add test for packaging weight (#158750)
yushangdi Aug 9, 2025
10e3514
Remove tensorexpr tests (#158928)
cyyever Aug 9, 2025
e07c52b
[dynamo] Improve support for itertools.product (#159693)
rtimpe Aug 8, 2025
5ed4f91
[dynamo] support itertools.permutations (#159694)
rtimpe Aug 8, 2025
0d88593
[audio hash update] update the pinned audio hash (#160153)
pytorchupdatebot Aug 9, 2025
303c614
[dynamo] Be consistent with UserMethodVariable source (#160155)
anijain2305 Aug 9, 2025
bcf23ec
[vllm hash update] update the pinned vllm hash (#160235)
pytorchupdatebot Aug 9, 2025
fb887c3
Add Sherlock and Zhengxu as codeowner for schema.py (#160233)
SherlockNoMad Aug 9, 2025
4183d4f
Make user defined Triton kernels serializable for fx_graph_runnable (…
PaulZhang12 Aug 8, 2025
8047421
[Linter] Expanding the scope of detecting device-bias code. (#159949)
etaf Aug 8, 2025
2f4c222
Revert "Make user defined Triton kernels serializable for fx_graph_ru…
pytorchmergebot Aug 9, 2025
01f66d0
Remove outdated CMAKE_CUDA_COMPILER_VERSION branch (#160075)
cyyever Aug 9, 2025
2971231
[fx][pass] Support converting a float32 tensor to a scalar in FX trac…
thenumberouscode Aug 9, 2025
db78943
Fix get_free_symbol_uses for several nodes. (#160134)
laithsakka Aug 9, 2025
f0980fc
[inductor] turn on windows inductor UTs (#160161)
xuhancn Aug 9, 2025
df55ec7
[OpInfo][BE] Better inputs for addmm (#160234)
malfet Aug 9, 2025
d3d359d
Revert "Fix get_free_symbol_uses for several nodes. (#160134)"
pytorchmergebot Aug 10, 2025
5dddcd5
Correctly copy self.module_stack in ModuleStackTracer (#159956)
ezyang Aug 6, 2025
af10f1f
Fix requires_cuda to requires_cuda_and_triton (#160222)
ghostspiders Aug 10, 2025
c9671dc
Delete Python reference implementation from torchdim, as it is untest…
ezyang Aug 10, 2025
3ac86e7
Add Alban and Piotr to list of maintainers (#160187)
atalman Aug 10, 2025
a84b60c
[MPS] Sparse coalesce more dtypes to match cpu (#160254)
Isalia20 Aug 10, 2025
0e3e377
[inductor] fix CompiledArtifact.load path on Windows. (#160268)
xuhancn Aug 10, 2025
7ae0629
Revert "[inductor] turn on windows inductor UTs (#160161)"
pytorchmergebot Aug 10, 2025
d678674
[inductor] slow test some Windows UTs. (#160267)
xuhancn Aug 10, 2025
05c19d1
[Inductor] Add back the revert part (#160054)
andyanwang Aug 10, 2025
4416433
[inductor] turn on windows inductor UTs (#160161)
xuhancn Aug 10, 2025
b602ea9
Revert "[inductor] turn on windows inductor UTs (#160161)"
pytorchmergebot Aug 11, 2025
842cc77
[MPS] Extend addmm to integral types (#160270)
malfet Aug 10, 2025
e7152ff
[inductor] fix some windows inductor UTs (#160292)
xuhancn Aug 11, 2025
d8cb3db
Add unsigned support to `IValue` (#160102)
malfet Aug 11, 2025
8088cfa
Add type assert for tensor_meta, based on real bug in autoparallel. (…
ezyang Aug 10, 2025
8ae4d26
Tidy torch/csrc/jit/passes/onnx code (#160262)
cyyever Aug 11, 2025
dc0d18e
[CUDA] Remove the uncessary CUDA_GUARD (#160249)
fffrog Aug 9, 2025
334b38c
Fix typo in README.md (#160160)
bjtuwjx Aug 11, 2025
ff0d56d
[Inductor] [Triton] Enable Configuration warmup/rep iterations when b…
njriasan Aug 11, 2025
1c2cba1
[FR] Add stack_id and an optional print of stack_id to stack_trace ma…
fduwjj Aug 8, 2025
ecea811
Fix clang builds by adding headers (#160252)
tanmay-sinha Aug 11, 2025
cf4964b
Remove unnecessary CMake checks for glog (#158185)
cyyever Aug 11, 2025
05029ad
[xla hash update] update the pinned xla hash (#160306)
pytorchupdatebot Aug 11, 2025
2259dbe
Update slow tests (#158222)
pytorchupdatebot Aug 11, 2025
c184cb3
[submodule] Bump fbgemm to latest (#158210)
cyyever Aug 11, 2025
515cb70
[inductor] normalize_path_separator for test_different_file_paths_loc…
xuhancn Aug 11, 2025
80cca83
[inductor] Skip some AOTI UTs on Windows. (#160287)
xuhancn Aug 11, 2025
68a4b4b
[codemod] Fix unreachable-break issue in caffe2/c10/cuda/CUDAFunction…
r-barnes Aug 11, 2025
ca7315c
[Graph Partition] Pass all OSS unit tests (#154667)
BoyuanFeng Aug 11, 2025
9ccd0f5
Fix unbacked symint and memory leak in inductor memory planning (#159…
yushangdi Aug 11, 2025
d0e2240
[triton_heuristics] Optimize the triton launcher in pt2 (#160000)
davidberard98 Aug 11, 2025
d25c4f9
[MPS] Type-promote tensor-iterator common dtype (#160334)
malfet Aug 11, 2025
c8205cb
[autograd] match 0-dim gradients device type regardless of subclassne…
xmfan Aug 9, 2025
76a0609
port distributed pipeline test files for Intel GPU (#159033)
wincent8 Aug 11, 2025
c3dc8dc
159965 is merged, no need to patch it in (#160275)
ezyang Aug 10, 2025
9eedd2a
[PGO] no counterfactual suggestions for dynamic allowlist (#160231)
pianpwk Aug 11, 2025
09381f5
Revert "[Graph Partition] Pass all OSS unit tests (#154667)"
pytorchmergebot Aug 11, 2025
b149c72
Revert "port distributed pipeline test files for Intel GPU (#159033)"
pytorchmergebot Aug 11, 2025
cf0a0dc
Make user defined Triton kernels serializable for fx_graph_runnable (…
PaulZhang12 Aug 11, 2025
fc80f68
Fix collective schedule logging and runtime tests (#160260)
skarjala Aug 11, 2025
7d2ec70
Fix MPS autocast for ConvTranspose3d (#160345)
malfet Aug 11, 2025
5a40c57
[MTIA] Implement isAvailable() for MTIA hooks (#160304)
PatriceVignola Aug 11, 2025
fc25c68
[hop][exc] make UncapturedHigherOrderOpError print user code and avoi…
ydwu4 Aug 11, 2025
99bc2f9
Update export/schema.py (#160220)
SherlockNoMad Aug 11, 2025
3626ba7
[FlexAttention] Swap from and to & for new triton (#160227)
drisspg Aug 11, 2025
e63c2b2
[PP] Initialize P2P communicators on first step (#160210)
H-Huang Aug 11, 2025
ee89cc7
[ROCm][Windows] Fix LoadHIP handling of environment variable paths on…
ScottTodd Aug 12, 2025
cae2b5e
[ROCm][Windows] Enable USE_ROCM, disable USE_RCCL on Windows. (#159079)
ScottTodd Aug 12, 2025
0d40ff3
[inductor] fix test_different_file_paths_local_pgo on Windows. (#160382)
xuhancn Aug 12, 2025
b90feea
[BE][cutlass backend] Fix subproc addmm tests (#160295)
henrylhtsang Aug 11, 2025
f3f159f
[BE][cutlass backend] Reduce severity of log message for no cutlass c…
henrylhtsang Aug 11, 2025
7a974a8
[ROCm] Fix resource_strings.h (#159996)
rraminen Aug 12, 2025
95210cc
[BE] Isolate pre-push hook dependencies in dedicated virtual environm…
ZainRizvi Aug 12, 2025
be53f60
fix retaining multimem in symmetric memory (#160343)
ngimel Aug 12, 2025
eed9dbf
[ROCm] Add torch/_rocm_init.py to .gitignore. (#159806)
ScottTodd Aug 12, 2025
bfc873d
[ROCm][Windows] Revert copying hipblaslt and rocblas dirs. (#159083)
ScottTodd Aug 12, 2025
32e5e2f
[vllm hash update] update the pinned vllm hash (#160259)
pytorchupdatebot Aug 12, 2025
10bc36f
Get tensor subclasses and torch.library.triton_op to dispatch correct…
zou3519 Aug 12, 2025
edaa151
[CI] Move CUDA tests to trunk workflow (#160379)
malfet Aug 11, 2025
5f1010f
[Graph Partition] Pass all OSS unit tests (#154667)
BoyuanFeng Aug 12, 2025
0f3b10b
[audio hash update] update the pinned audio hash (#160384)
pytorchupdatebot Aug 12, 2025
8d3d1c8
[dynamo] fixes to propagate tag safeness (#159807)
anijain2305 Aug 12, 2025
01bcf9a
Bump transformers pin (#159291)
anijain2305 Aug 12, 2025
9a0f7a3
[retry-land][pytorch][dynamo_compile] Log stack_trace to dynamo_compi…
jovianjaison Aug 12, 2025
fea7e9d
extract shape in _view_has_unbacked_input (#160255)
laithsakka Aug 12, 2025
b9003ed
Dynamo Deep Dive Documentation Fix (#158860)
morrison-turnansky Aug 12, 2025
f990490
Add `label_smoothing` param in `nn.BCELoss` and `nn.BCEWithLogitsLoss…
zeshengzong Aug 12, 2025
4d5b3f2
[dynamo][guards] Install dict watchers for recrusive dict tag optimiz…
anijain2305 Aug 12, 2025
f33ce40
[bucketing] Bucket only adjacent collectives to prevent reordering (#…
IvanKobzarev Aug 7, 2025
7fbc228
Update triton xpu commit to support python 3.14 (#160183)
chuanqi129 Aug 12, 2025
a288b15
[CI] Reduce XPU Windows build time (#159763)
chuanqi129 Aug 12, 2025
9708fcf
Account for triton kernel source code hidden in custom ops properly i…
jamesjwu Aug 10, 2025
b7db866
Fix Tensor illustration, use permalinks for image embedding in Readme…
atalman Aug 12, 2025
b219ca2
Revert "Update triton xpu commit to support python 3.14 (#160183)"
pytorchmergebot Aug 12, 2025
9d37c96
[ROCm][CI] use new benchmark image for dynamo (#160421)
jeffdaily Aug 12, 2025
f7b2f33
Revert "[triton_heuristics] Optimize the triton launcher in pt2 (#160…
pytorchmergebot Aug 12, 2025
a7abf57
[ROCm] Support large inputs for coalesceValuesKernel (#158281)
xinyazhang Aug 12, 2025
94b91a8
[redone][pytorch] Moving torch.compile worker process logs to a dedic…
jovianjaison Aug 12, 2025
1f4057c
[inductor] remove no_x_dim (#159810)
davidberard98 Aug 11, 2025
ee9f8ba
[ROCm] Use opportunistic fastatomics based on hueristics (#159430)
jerrymannil Aug 12, 2025
3cec82a
Ensure outer aliasing on DTensor matches inner aliasing (#158954)
ezyang Aug 12, 2025
f341077
Revert "[ROCm] Support large inputs for coalesceValuesKernel (#158281)"
pytorchmergebot Aug 12, 2025
9903ca4
[cuDNN][64-bit indexing] update conv depthwise 64bit indexing dispatc…
eqy Aug 12, 2025
2d0cdee
move thread-local capture mode guard to include work.isStarted (#160398)
ngimel Aug 12, 2025
89654db
[inductor] fix triton bucketize mask propagation (#159961)
v0i0 Aug 12, 2025
7e91394
Support NUMA Binding for Callable Entrypoints (#160163)
pdesupinski Aug 12, 2025
8e6a313
Add ownership token when needed on GradientEdge (#160098)
albanD Aug 12, 2025
f95b58c
Remove usage of fsspec in HF consolidation script (#159392)
ankitageorge Aug 12, 2025
a354fa9
added class or module info for functions blocked by weight-only load …
zero000064 Aug 12, 2025
5a9c4cf
[Fix XPU CI][Inductor UT] Fix test cases broken by community. (#160403)
etaf Aug 12, 2025
b459689
[DTensor] Registers sharding rule for rms_norm (#159692)
AaronWang04 Aug 12, 2025
c24ca7f
[FSDP][Collectives] skipping allgather when world size is 1 (#160135)
anshul-si Aug 12, 2025
f27232a
[ROCm] Limit number of values per thread for reductions on three dime…
doru1004 Aug 12, 2025
655137b
Update torch::stable::Tensor() default constructor (#159507)
mikaylagawarecki Aug 12, 2025
4d419a7
Add pad and narrow to torch/csrc/stable/ops.h (#159328)
mikaylagawarecki Aug 12, 2025
f8f0414
fix cpp builder to avoid missing-source compile error (#160354)
izaitsevfb Aug 12, 2025
78a2fe1
[TorchScript] thread-safe ErrorReport::CallStack (#160386)
davidberard98 Aug 12, 2025
cbffde7
Factor out the strings to templates for better editor integration (#1…
drisspg Aug 12, 2025
101276f
[BE] Save attributes for CppCompileError for pickleing (#160294)
henrylhtsang Aug 12, 2025
16d1544
Fullgraph graph capture with dynamo. (#159749)
zhxchen17 Aug 12, 2025
2e4e5ab
[MPS] Add mps keys to `indices` and `values` ops (#160223)
Isalia20 Aug 12, 2025
5737372
[CI] Switch ROCm MI300 GitHub Actions workflows from 2-GPU to 1-GPU r…
deedongala Aug 12, 2025
0d71ca2
[EZ] Replace `pytorch-labs` with `meta-pytorch` (#160459)
ZainRizvi Aug 12, 2025
b1f4354
[c10d] Error out the case when registering symmetric memory without e…
fduwjj Aug 12, 2025
8d1cf52
[EZ][BE] Remove unused `conda-env-macOS-ARM64` (#160477)
malfet Aug 12, 2025
3209996
[EZ] Delete CircleCI case (#160479)
malfet Aug 12, 2025
69a0a9a
[Inductor][Triton] Pass GPUTarget param to updated make_ir function (…
NikhilAPatel Aug 13, 2025
f15ada5
Enable output padding when only outermost dim is dynamic (#159404)
nandesuka Aug 13, 2025
e992eaa
support optim cases
Chao1Han Jul 21, 2025
62eb2ca
blacklist gloo
Chao1Han Jul 29, 2025
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
16 changes: 4 additions & 12 deletions .ci/aarch64_linux/build_aarch64_wheel.py
Original file line number Diff line number Diff line change
Expand Up @@ -438,9 +438,7 @@ def build_torchvision(
)
build_vars += f"BUILD_VERSION={version}.dev{build_date}"
elif build_version is not None:
build_vars += (
f"BUILD_VERSION={build_version} PYTORCH_VERSION={branch[1:].split('-')[0]}"
)
build_vars += f"BUILD_VERSION={build_version} PYTORCH_VERSION={branch[1:].split('-', maxsplit=1)[0]}"
if host.using_docker():
build_vars += " CMAKE_SHARED_LINKER_FLAGS=-Wl,-z,max-page-size=0x10000"

Expand Down Expand Up @@ -495,9 +493,7 @@ def build_torchdata(
)
build_vars += f"BUILD_VERSION={version}.dev{build_date}"
elif build_version is not None:
build_vars += (
f"BUILD_VERSION={build_version} PYTORCH_VERSION={branch[1:].split('-')[0]}"
)
build_vars += f"BUILD_VERSION={build_version} PYTORCH_VERSION={branch[1:].split('-', maxsplit=1)[0]}"
if host.using_docker():
build_vars += " CMAKE_SHARED_LINKER_FLAGS=-Wl,-z,max-page-size=0x10000"

Expand Down Expand Up @@ -553,9 +549,7 @@ def build_torchtext(
)
build_vars += f"BUILD_VERSION={version}.dev{build_date}"
elif build_version is not None:
build_vars += (
f"BUILD_VERSION={build_version} PYTORCH_VERSION={branch[1:].split('-')[0]}"
)
build_vars += f"BUILD_VERSION={build_version} PYTORCH_VERSION={branch[1:].split('-', maxsplit=1)[0]}"
if host.using_docker():
build_vars += " CMAKE_SHARED_LINKER_FLAGS=-Wl,-z,max-page-size=0x10000"

Expand Down Expand Up @@ -613,9 +607,7 @@ def build_torchaudio(
)
build_vars += f"BUILD_VERSION={version}.dev{build_date}"
elif build_version is not None:
build_vars += (
f"BUILD_VERSION={build_version} PYTORCH_VERSION={branch[1:].split('-')[0]}"
)
build_vars += f"BUILD_VERSION={build_version} PYTORCH_VERSION={branch[1:].split('-', maxsplit=1)[0]}"
if host.using_docker():
build_vars += " CMAKE_SHARED_LINKER_FLAGS=-Wl,-z,max-page-size=0x10000"

Expand Down
62 changes: 4 additions & 58 deletions .ci/docker/build.sh
Original file line number Diff line number Diff line change
Expand Up @@ -144,16 +144,6 @@ case "$tag" in
TRITON=yes
INDUCTOR_BENCHMARKS=yes
;;
pytorch-linux-jammy-cuda12.6-cudnn9-py3-gcc9)
CUDA_VERSION=12.6.3
ANACONDA_PYTHON_VERSION=3.10
GCC_VERSION=9
VISION=yes
KATEX=yes
UCX_COMMIT=${_UCX_COMMIT}
UCC_COMMIT=${_UCC_COMMIT}
TRITON=yes
;;
pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm)
CUDA_VERSION=12.8.1
ANACONDA_PYTHON_VERSION=3.12
Expand All @@ -164,39 +154,6 @@ case "$tag" in
UCC_COMMIT=${_UCC_COMMIT}
TRITON=yes
;;
pytorch-linux-jammy-cuda12.6-cudnn9-py3-gcc9-inductor-benchmarks)
CUDA_VERSION=12.6
ANACONDA_PYTHON_VERSION=3.10
GCC_VERSION=9
VISION=yes
KATEX=yes
UCX_COMMIT=${_UCX_COMMIT}
UCC_COMMIT=${_UCC_COMMIT}
TRITON=yes
INDUCTOR_BENCHMARKS=yes
;;
pytorch-linux-jammy-cuda12.6-cudnn9-py3.12-gcc9-inductor-benchmarks)
CUDA_VERSION=12.6
ANACONDA_PYTHON_VERSION=3.12
GCC_VERSION=9
VISION=yes
KATEX=yes
UCX_COMMIT=${_UCX_COMMIT}
UCC_COMMIT=${_UCC_COMMIT}
TRITON=yes
INDUCTOR_BENCHMARKS=yes
;;
pytorch-linux-jammy-cuda12.6-cudnn9-py3.13-gcc9-inductor-benchmarks)
CUDA_VERSION=12.6
ANACONDA_PYTHON_VERSION=3.13
GCC_VERSION=9
VISION=yes
KATEX=yes
UCX_COMMIT=${_UCX_COMMIT}
UCC_COMMIT=${_UCC_COMMIT}
TRITON=yes
INDUCTOR_BENCHMARKS=yes
;;
pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc9)
CUDA_VERSION=12.8.1
ANACONDA_PYTHON_VERSION=3.10
Expand All @@ -219,19 +176,7 @@ case "$tag" in
VISION=yes
TRITON=yes
;;
pytorch-linux-jammy-py3.11-clang12)
ANACONDA_PYTHON_VERSION=3.11
CLANG_VERSION=12
VISION=yes
TRITON=yes
;;
pytorch-linux-jammy-py3.9-gcc9)
ANACONDA_PYTHON_VERSION=3.9
GCC_VERSION=9
VISION=yes
TRITON=yes
;;
pytorch-linux-jammy-rocm-n-py3 | pytorch-linux-noble-rocm-n-py3)
pytorch-linux-jammy-rocm-n-py3 | pytorch-linux-jammy-rocm-n-py3-benchmarks | pytorch-linux-noble-rocm-n-py3)
if [[ $tag =~ "jammy" ]]; then
ANACONDA_PYTHON_VERSION=3.10
else
Expand All @@ -245,7 +190,9 @@ case "$tag" in
KATEX=yes
UCX_COMMIT=${_UCX_COMMIT}
UCC_COMMIT=${_UCC_COMMIT}
INDUCTOR_BENCHMARKS=yes
if [[ $tag =~ "benchmarks" ]]; then
INDUCTOR_BENCHMARKS=yes
fi
;;
pytorch-linux-noble-rocm-alpha-py3)
ANACONDA_PYTHON_VERSION=3.12
Expand All @@ -257,7 +204,6 @@ case "$tag" in
KATEX=yes
UCX_COMMIT=${_UCX_COMMIT}
UCC_COMMIT=${_UCC_COMMIT}
INDUCTOR_BENCHMARKS=yes
PYTORCH_ROCM_ARCH="gfx90a;gfx942;gfx950"
;;
pytorch-linux-jammy-xpu-2025.0-py3)
Expand Down
2 changes: 1 addition & 1 deletion .ci/docker/ci_commit_pins/huggingface.txt
Original file line number Diff line number Diff line change
@@ -1 +1 @@
243e186efbf7fb93328dd6b34927a4e8c8f24395
v4.54.0
2 changes: 1 addition & 1 deletion .ci/docker/ci_commit_pins/triton.txt
Original file line number Diff line number Diff line change
@@ -1 +1 @@
11ec6354315768a85da41032535e3b7b99c5f706
f7888497a1eb9e98d4c07537f0d0bcfe180d1363
5 changes: 3 additions & 2 deletions .ci/docker/common/install_cpython.sh
Original file line number Diff line number Diff line change
Expand Up @@ -66,8 +66,9 @@ function do_cpython_build {
ln -s pip3 ${prefix}/bin/pip
fi
# install setuptools since python 3.12 is required to use distutils
${prefix}/bin/pip install wheel==0.45.1 setuptools==80.9.0
local abi_tag=$(${prefix}/bin/python -c "from wheel.pep425tags import get_abbr_impl, get_impl_ver, get_abi_tag; print('{0}{1}-{2}'.format(get_abbr_impl(), get_impl_ver(), get_abi_tag()))")
# packaging is needed to create symlink since wheel no longer provides needed information
${prefix}/bin/pip install packaging==25.0 wheel==0.45.1 setuptools==80.9.0
local abi_tag=$(${prefix}/bin/python -c "from packaging.tags import interpreter_name, interpreter_version; import sysconfig ; from sysconfig import get_config_var; print('{0}{1}-{0}{1}{2}'.format(interpreter_name(), interpreter_version(), 't' if sysconfig.get_config_var('Py_GIL_DISABLED') else ''))")
ln -sf ${prefix} /opt/python/${abi_tag}
}

Expand Down
4 changes: 2 additions & 2 deletions .ci/docker/common/install_cuda.sh
Original file line number Diff line number Diff line change
Expand Up @@ -68,8 +68,8 @@ function install_nvshmem {
# download, unpack, install
wget -q "${url}"
tar xf "${filename}.tar.gz"
cp -a "libnvshmem/include/"* /usr/local/include/
cp -a "libnvshmem/lib/"* /usr/local/lib/
cp -a "libnvshmem/include/"* /usr/local/cuda/include/
cp -a "libnvshmem/lib/"* /usr/local/cuda/lib64/

# cleanup
cd ..
Expand Down
30 changes: 28 additions & 2 deletions .ci/docker/common/install_inductor_benchmark_deps.sh
Original file line number Diff line number Diff line change
Expand Up @@ -15,11 +15,37 @@ function install_timm() {
commit=$(get_pinned_commit timm)

pip_install "git+https://github.com/huggingface/pytorch-image-models@${commit}"
# Clean up
conda_run pip uninstall -y torch torchvision triton
}

function install_torchbench() {
local commit
commit=$(get_pinned_commit torchbench)
git clone https://github.com/pytorch/benchmark torchbench
pushd torchbench
git checkout "$commit"

python install.py --continue_on_fail

# soxr comes from https://github.com/huggingface/transformers/pull/39429
pip install transformers==4.54.0 soxr==0.5.0

echo "Print all dependencies after TorchBench is installed"
python -mpip freeze
popd

chown -R jenkins torchbench
chown -R jenkins /opt/conda
}

# Pango is needed for weasyprint which is needed for doctr
conda_install pango

# Stable packages are ok here, just to satisfy TorchBench check
pip_install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu128

install_torchbench
install_huggingface
install_timm

# Clean up
conda_run pip uninstall -y torch torchvision torchaudio triton torchao
2 changes: 1 addition & 1 deletion .ci/docker/common/install_triton.sh
Original file line number Diff line number Diff line change
Expand Up @@ -103,5 +103,5 @@ fi
# It depends on torch and triton. We don't want to install
# triton and torch from production on Docker CI images
if [[ "$ANACONDA_PYTHON_VERSION" != 3.9* ]]; then
pip_install helion==0.0.10 --no-deps
pip_install helion --no-deps
fi
41 changes: 25 additions & 16 deletions .ci/docker/common/install_xpu.sh
Original file line number Diff line number Diff line change
Expand Up @@ -34,18 +34,27 @@ function install_ubuntu() {

# The xpu-smi packages
apt-get install -y flex bison xpu-smi
# Compute and Media Runtimes
apt-get install -y \
intel-opencl-icd intel-level-zero-gpu level-zero \
intel-media-va-driver-non-free libmfx1 libmfxgen1 libvpl2 \
libegl-mesa0 libegl1-mesa libegl1-mesa-dev libgbm1 libgl1-mesa-dev libgl1-mesa-dri \
libglapi-mesa libgles2-mesa-dev libglx-mesa0 libigdgmm12 libxatracker2 mesa-va-drivers \
mesa-vdpau-drivers mesa-vulkan-drivers va-driver-all vainfo hwinfo clinfo
if [[ "${XPU_DRIVER_TYPE,,}" == "rolling" ]]; then
apt-get install -y intel-ocloc

if [[ "${XPU_DRIVER_TYPE,,}" == "lts" ]]; then
# Compute and Media Runtimes
apt-get install -y \
intel-opencl-icd intel-level-zero-gpu level-zero \
intel-media-va-driver-non-free libmfx1 libmfxgen1 libvpl2 \
libegl-mesa0 libegl1-mesa libegl1-mesa-dev libgbm1 libgl1-mesa-dev libgl1-mesa-dri \
libglapi-mesa libgles2-mesa-dev libglx-mesa0 libigdgmm12 libxatracker2 mesa-va-drivers \
mesa-vdpau-drivers mesa-vulkan-drivers va-driver-all vainfo hwinfo clinfo
# Development Packages
apt-get install -y libigc-dev intel-igc-cm libigdfcl-dev libigfxcmrt-dev level-zero-dev
else # rolling driver
apt-get install -y \
intel-opencl-icd libze-intel-gpu1 libze1 \
intel-media-va-driver-non-free libmfx-gen1 libvpl2 \
libegl-mesa0 libegl1-mesa libegl1-mesa-dev libgbm1 libgl1-mesa-dev libgl1-mesa-dri \
libglapi-mesa libglx-mesa0 libigdgmm12 libxatracker2 mesa-va-drivers \
mesa-vdpau-drivers mesa-vulkan-drivers va-driver-all vainfo hwinfo clinfo intel-ocloc
apt-get install -y libigc-dev intel-igc-cm libigdfcl-dev libigfxcmrt-dev libze-dev
fi
# Development Packages
apt-get install -y libigc-dev intel-igc-cm libigdfcl-dev libigfxcmrt-dev level-zero-dev

# Install Intel Support Packages
apt-get install -y ${XPU_PACKAGES}

Expand Down Expand Up @@ -130,11 +139,11 @@ function install_sles() {

}

# Default use GPU driver LTS releases
XPU_DRIVER_VERSION="/lts/2350"
if [[ "${XPU_DRIVER_TYPE,,}" == "rolling" ]]; then
# Use GPU driver rolling releases
XPU_DRIVER_VERSION=""
# Default use GPU driver rolling releases
XPU_DRIVER_VERSION=""
if [[ "${XPU_DRIVER_TYPE,,}" == "lts" ]]; then
# Use GPU driver LTS releases
XPU_DRIVER_VERSION="/lts/2350"
fi

# Default use Intel® oneAPI Deep Learning Essentials 2025.0
Expand Down
17 changes: 9 additions & 8 deletions .ci/docker/requirements-ci.txt
Original file line number Diff line number Diff line change
Expand Up @@ -63,11 +63,12 @@ lark==0.12.0
#Pinned versions: 0.12.0
#test that import:

librosa>=0.6.2 ; python_version < "3.11"
librosa==0.10.2 ; python_version == "3.12"
librosa>=0.6.2 ; python_version < "3.11" and platform_machine != "s390x"
librosa==0.10.2 ; python_version == "3.12" and platform_machine != "s390x"
#Description: A python package for music and audio analysis
#Pinned versions: >=0.6.2
#test that import: test_spectral_ops.py
#librosa depends on numba; disable it for s390x while numba is disabled too

#mkl #this breaks linux-bionic-rocm4.5-py3.7
#Description: Intel oneAPI Math Kernel Library
Expand Down Expand Up @@ -110,14 +111,15 @@ ninja==1.11.1.3
#Pinned versions: 1.11.1.3
#test that import: run_test.py, test_cpp_extensions_aot.py,test_determination.py

numba==0.49.0 ; python_version < "3.9"
numba==0.55.2 ; python_version == "3.9"
numba==0.55.2 ; python_version == "3.10"
numba==0.60.0 ; python_version == "3.12"
numba==0.49.0 ; python_version < "3.9" and platform_machine != "s390x"
numba==0.55.2 ; python_version == "3.9" and platform_machine != "s390x"
numba==0.55.2 ; python_version == "3.10" and platform_machine != "s390x"
numba==0.60.0 ; python_version == "3.12" and platform_machine != "s390x"
#Description: Just-In-Time Compiler for Numerical Functions
#Pinned versions: 0.54.1, 0.49.0, <=0.49.1
#test that import: test_numba_integration.py
#For numba issue see https://github.com/pytorch/pytorch/issues/51511
#Need release > 0.61.2 for s390x due to https://github.com/numba/numba/pull/10073

#numpy
#Description: Provides N-dimensional arrays and linear algebra
Expand Down Expand Up @@ -307,7 +309,7 @@ pytest-cpp==2.3.0
#Pinned versions: 2.3.0
#test that import:

z3-solver==4.15.1.0
z3-solver==4.15.1.0 ; platform_machine != "s390x"
#Description: The Z3 Theorem Prover Project
#Pinned versions:
#test that import:
Expand Down Expand Up @@ -361,7 +363,6 @@ pwlf==2.2.1
#Pinned versions: 2.2.1
#test that import: test_sac_estimator.py


# To build PyTorch itself
pyyaml
pyzstd
Expand Down
4 changes: 2 additions & 2 deletions .ci/docker/requirements-docs.txt
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
sphinx==5.3.0
#Description: This is used to generate PyTorch docs
#Pinned versions: 5.3.0
-e git+https://github.com/pytorch/pytorch_sphinx_theme.git@pytorch_sphinx_theme2#egg=pytorch_sphinx_theme2
-e git+https://github.com/pytorch/pytorch_sphinx_theme.git@722b7e6f9ca512fcc526ad07d62b3d28c50bb6cd#egg=pytorch_sphinx_theme2

# TODO: sphinxcontrib.katex 0.9.0 adds a local KaTeX server to speed up pre-rendering
# but it doesn't seem to work and hangs around idly. The initial thought that it is probably
Expand Down Expand Up @@ -50,7 +50,7 @@ IPython==8.12.0
#Pinned versions: 8.12.0

myst-nb==0.17.2
#Description: This is used to generate PyTorch functorch and torch.compile docs
#Description: This is used to generate PyTorch functorch and torch.compile docs.
#Pinned versions: 0.17.2

# The following are required to build torch.distributed.elastic.rendezvous.etcd* docs
Expand Down
3 changes: 2 additions & 1 deletion .ci/docker/ubuntu-rocm/Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -98,8 +98,9 @@ COPY ./common/install_inductor_benchmark_deps.sh install_inductor_benchmark_deps
COPY ./common/common_utils.sh common_utils.sh
COPY ci_commit_pins/huggingface.txt huggingface.txt
COPY ci_commit_pins/timm.txt timm.txt
COPY ci_commit_pins/torchbench.txt torchbench.txt
RUN if [ -n "${INDUCTOR_BENCHMARKS}" ]; then bash ./install_inductor_benchmark_deps.sh; fi
RUN rm install_inductor_benchmark_deps.sh common_utils.sh timm.txt huggingface.txt
RUN rm install_inductor_benchmark_deps.sh common_utils.sh timm.txt huggingface.txt torchbench.txt

# (optional) Install non-default Ninja version
ARG NINJA_VERSION
Expand Down
3 changes: 2 additions & 1 deletion .ci/docker/ubuntu/Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -98,8 +98,9 @@ COPY ./common/install_inductor_benchmark_deps.sh install_inductor_benchmark_deps
COPY ./common/common_utils.sh common_utils.sh
COPY ci_commit_pins/huggingface.txt huggingface.txt
COPY ci_commit_pins/timm.txt timm.txt
COPY ci_commit_pins/torchbench.txt torchbench.txt
RUN if [ -n "${INDUCTOR_BENCHMARKS}" ]; then bash ./install_inductor_benchmark_deps.sh; fi
RUN rm install_inductor_benchmark_deps.sh common_utils.sh timm.txt huggingface.txt
RUN rm install_inductor_benchmark_deps.sh common_utils.sh timm.txt huggingface.txt torchbench.txt

ARG TRITON
ARG TRITON_CPU
Expand Down
4 changes: 0 additions & 4 deletions .ci/manywheel/build.sh
Original file line number Diff line number Diff line change
Expand Up @@ -5,10 +5,6 @@ set -ex
SCRIPTPATH="$( cd "$( dirname "${BASH_SOURCE[0]}" )" >/dev/null 2>&1 && pwd )"

case "${GPU_ARCH_TYPE:-BLANK}" in
BLANK)
# Legacy behavior for CircleCI
bash "${SCRIPTPATH}/build_cuda.sh"
;;
cuda)
bash "${SCRIPTPATH}/build_cuda.sh"
;;
Expand Down
Loading