torch-rb 0.17.0 → 0.18.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/CHANGELOG.md +9 -0
- data/README.md +14 -18
- data/codegen/native_functions.yaml +230 -151
- data/ext/torch/utils.h +1 -1
- data/lib/torch/nn/module.rb +24 -4
- data/lib/torch/tensor.rb +8 -3
- data/lib/torch/version.rb +1 -1
- metadata +5 -5
@@ -338,7 +338,7 @@
|
|
338
338
|
dispatch:
|
339
339
|
CompositeExplicitAutograd: abs
|
340
340
|
SparseCPU, SparseCUDA: abs_sparse
|
341
|
-
SparseCsrCPU, SparseCsrCUDA: abs_sparse_csr
|
341
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: abs_sparse_csr
|
342
342
|
NestedTensorCPU, NestedTensorCUDA: NestedTensor_abs
|
343
343
|
tags: [core, pointwise]
|
344
344
|
|
@@ -348,7 +348,7 @@
|
|
348
348
|
dispatch:
|
349
349
|
CompositeExplicitAutograd: abs_
|
350
350
|
SparseCPU, SparseCUDA: abs_sparse_
|
351
|
-
SparseCsrCPU, SparseCsrCUDA: abs_sparse_csr_
|
351
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: abs_sparse_csr_
|
352
352
|
NestedTensorCPU, NestedTensorCUDA: NestedTensor_abs_
|
353
353
|
|
354
354
|
- func: abs.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)
|
@@ -357,7 +357,7 @@
|
|
357
357
|
CPU, CUDA: abs_out
|
358
358
|
MPS: abs_out_mps
|
359
359
|
SparseCPU, SparseCUDA: abs_sparse_out
|
360
|
-
SparseCsrCPU, SparseCsrCUDA: abs_sparse_csr_out
|
360
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: abs_sparse_csr_out
|
361
361
|
tags: pointwise
|
362
362
|
|
363
363
|
# Note [Adding an alias]
|
@@ -400,14 +400,14 @@
|
|
400
400
|
variants: function, method
|
401
401
|
dispatch:
|
402
402
|
CPU, CUDA: angle
|
403
|
-
SparseCsrCPU, SparseCsrCUDA: angle_sparse_csr
|
403
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: angle_sparse_csr
|
404
404
|
tags: pointwise
|
405
405
|
|
406
406
|
- func: angle.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)
|
407
407
|
device_check: NoCheck # TensorIterator
|
408
408
|
dispatch:
|
409
409
|
CPU, CUDA: angle_out
|
410
|
-
SparseCsrCPU, SparseCsrCUDA: angle_sparse_csr_out
|
410
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: angle_sparse_csr_out
|
411
411
|
tags: pointwise
|
412
412
|
|
413
413
|
- func: view_as_real(Tensor(a) self) -> Tensor(a)
|
@@ -425,7 +425,7 @@
|
|
425
425
|
structured_delegate: sgn.out
|
426
426
|
dispatch:
|
427
427
|
SparseCPU, SparseCUDA: sgn_sparse
|
428
|
-
SparseCsrCPU, SparseCsrCUDA: sgn_sparse_csr
|
428
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: sgn_sparse_csr
|
429
429
|
NestedTensorCPU, NestedTensorCUDA: NestedTensor_sgn
|
430
430
|
tags: pointwise
|
431
431
|
|
@@ -434,7 +434,7 @@
|
|
434
434
|
structured_delegate: sgn.out
|
435
435
|
dispatch:
|
436
436
|
SparseCPU, SparseCUDA: sgn_sparse_
|
437
|
-
SparseCsrCPU, SparseCsrCUDA: sgn_sparse_csr_
|
437
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: sgn_sparse_csr_
|
438
438
|
NestedTensorCPU, NestedTensorCUDA: NestedTensor_sgn_
|
439
439
|
tags: pointwise
|
440
440
|
|
@@ -445,7 +445,7 @@
|
|
445
445
|
CPU, CUDA: sgn_out
|
446
446
|
MPS: sgn_out_mps
|
447
447
|
SparseCPU, SparseCUDA: sgn_sparse_out
|
448
|
-
SparseCsrCPU, SparseCsrCUDA: sgn_sparse_csr_out
|
448
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: sgn_sparse_csr_out
|
449
449
|
tags: pointwise
|
450
450
|
|
451
451
|
- func: chalf(Tensor self, *, MemoryFormat? memory_format=None) -> Tensor
|
@@ -472,7 +472,7 @@
|
|
472
472
|
variants: function, method
|
473
473
|
dispatch:
|
474
474
|
CompositeExplicitAutograd: _conj_physical
|
475
|
-
SparseCsrCPU, SparseCsrCUDA: conj_physical_sparse_csr
|
475
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: conj_physical_sparse_csr
|
476
476
|
autogen: _conj_physical.out
|
477
477
|
|
478
478
|
- func: conj_physical(Tensor self) -> Tensor
|
@@ -484,14 +484,14 @@
|
|
484
484
|
CPU, CUDA: conj_physical_out
|
485
485
|
MPS: conj_physical_out_mps
|
486
486
|
SparseCPU, SparseCUDA: conj_physical_out_sparse
|
487
|
-
SparseCsrCPU, SparseCsrCUDA: conj_physical_sparse_csr_out
|
487
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: conj_physical_sparse_csr_out
|
488
488
|
tags: pointwise
|
489
489
|
|
490
490
|
- func: conj_physical_(Tensor(a!) self) -> Tensor(a!)
|
491
491
|
variants: function, method
|
492
492
|
dispatch:
|
493
493
|
CompositeExplicitAutograd: conj_physical_
|
494
|
-
SparseCsrCPU, SparseCsrCUDA: conj_physical_sparse_csr_
|
494
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: conj_physical_sparse_csr_
|
495
495
|
tags: pointwise
|
496
496
|
|
497
497
|
- func: resolve_conj(Tensor(a) self) -> Tensor(a)
|
@@ -537,9 +537,11 @@
|
|
537
537
|
|
538
538
|
- func: avg_pool1d(Tensor self, int[1] kernel_size, int[1] stride=[], int[1] padding=0, bool ceil_mode=False, bool count_include_pad=True) -> Tensor
|
539
539
|
tags: core
|
540
|
+
autogen: avg_pool1d.out
|
540
541
|
|
541
542
|
- func: adaptive_avg_pool1d(Tensor self, int[1] output_size) -> Tensor
|
542
543
|
tags: core
|
544
|
+
autogen: adaptive_avg_pool1d.out
|
543
545
|
|
544
546
|
# Return: (Tensor output, Tensor indices)
|
545
547
|
- func: adaptive_max_pool1d(Tensor self, int[1] output_size) -> (Tensor, Tensor)
|
@@ -694,6 +696,9 @@
|
|
694
696
|
device_check: NoCheck # TensorIterator
|
695
697
|
structured_delegate: all.out
|
696
698
|
variants: function, method
|
699
|
+
dispatch:
|
700
|
+
NestedTensorCPU, NestedTensorCUDA: NestedTensor_all
|
701
|
+
|
697
702
|
|
698
703
|
- func: all.dims(Tensor self, int[]? dim=None, bool keepdim=False) -> Tensor
|
699
704
|
device_check: NoCheck # TensorIterator
|
@@ -863,7 +868,7 @@
|
|
863
868
|
structured_delegate: asinh.out
|
864
869
|
dispatch:
|
865
870
|
SparseCPU, SparseCUDA: asinh_sparse
|
866
|
-
SparseCsrCPU, SparseCsrCUDA: asinh_sparse_csr
|
871
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: asinh_sparse_csr
|
867
872
|
tags: [core, pointwise]
|
868
873
|
|
869
874
|
- func: asinh_(Tensor(a!) self) -> Tensor(a!)
|
@@ -871,7 +876,7 @@
|
|
871
876
|
structured_delegate: asinh.out
|
872
877
|
dispatch:
|
873
878
|
SparseCPU, SparseCUDA: asinh_sparse_
|
874
|
-
SparseCsrCPU, SparseCsrCUDA: asinh_sparse_csr_
|
879
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: asinh_sparse_csr_
|
875
880
|
tags: pointwise
|
876
881
|
|
877
882
|
- func: asinh.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)
|
@@ -881,7 +886,7 @@
|
|
881
886
|
CPU, CUDA: asinh_out
|
882
887
|
MPS: asinh_out_mps
|
883
888
|
SparseCPU, SparseCUDA: asinh_sparse_out
|
884
|
-
SparseCsrCPU, SparseCsrCUDA: asinh_sparse_csr_out
|
889
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: asinh_sparse_csr_out
|
885
890
|
tags: pointwise
|
886
891
|
|
887
892
|
# arcsinh, alias for asinh
|
@@ -898,7 +903,7 @@
|
|
898
903
|
variants: function, method
|
899
904
|
dispatch:
|
900
905
|
SparseCPU, SparseCUDA: atanh_sparse
|
901
|
-
SparseCsrCPU, SparseCsrCUDA: atanh_sparse_csr
|
906
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: atanh_sparse_csr
|
902
907
|
tags: [core, pointwise]
|
903
908
|
|
904
909
|
- func: atanh_(Tensor(a!) self) -> Tensor(a!)
|
@@ -906,7 +911,7 @@
|
|
906
911
|
variants: function, method
|
907
912
|
dispatch:
|
908
913
|
SparseCPU, SparseCUDA: atanh_sparse_
|
909
|
-
SparseCsrCPU, SparseCsrCUDA: atanh_sparse_csr_
|
914
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: atanh_sparse_csr_
|
910
915
|
tags: pointwise
|
911
916
|
|
912
917
|
- func: atanh.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)
|
@@ -916,7 +921,7 @@
|
|
916
921
|
CPU, CUDA: atanh_out
|
917
922
|
MPS: atanh_out_mps
|
918
923
|
SparseCPU, SparseCUDA: atanh_sparse_out
|
919
|
-
SparseCsrCPU, SparseCsrCUDA: atanh_sparse_csr_out
|
924
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: atanh_sparse_csr_out
|
920
925
|
tags: pointwise
|
921
926
|
# arctanh, alias for atanh
|
922
927
|
|
@@ -954,7 +959,7 @@
|
|
954
959
|
structured_delegate: asin.out
|
955
960
|
dispatch:
|
956
961
|
SparseCPU, SparseCUDA: asin_sparse
|
957
|
-
SparseCsrCPU, SparseCsrCUDA: asin_sparse_csr
|
962
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: asin_sparse_csr
|
958
963
|
tags: [core, pointwise]
|
959
964
|
|
960
965
|
- func: asin_(Tensor(a!) self) -> Tensor(a!)
|
@@ -963,7 +968,7 @@
|
|
963
968
|
structured_delegate: asin.out
|
964
969
|
dispatch:
|
965
970
|
SparseCPU, SparseCUDA: asin_sparse_
|
966
|
-
SparseCsrCPU, SparseCsrCUDA: asin_sparse_csr_
|
971
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: asin_sparse_csr_
|
967
972
|
tags: pointwise
|
968
973
|
|
969
974
|
- func: asin.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)
|
@@ -974,7 +979,7 @@
|
|
974
979
|
CPU, CUDA: asin_out
|
975
980
|
MPS: asin_out_mps
|
976
981
|
SparseCPU, SparseCUDA: asin_sparse_out
|
977
|
-
SparseCsrCPU, SparseCsrCUDA: asin_sparse_csr_out
|
982
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: asin_sparse_csr_out
|
978
983
|
tags: pointwise
|
979
984
|
|
980
985
|
# arcsin, alias of asin
|
@@ -992,7 +997,7 @@
|
|
992
997
|
variants: function, method
|
993
998
|
dispatch:
|
994
999
|
SparseCPU, SparseCUDA: atan_sparse
|
995
|
-
SparseCsrCPU, SparseCsrCUDA: atan_sparse_csr
|
1000
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: atan_sparse_csr
|
996
1001
|
tags: [core, pointwise]
|
997
1002
|
|
998
1003
|
- func: atan_(Tensor(a!) self) -> Tensor(a!)
|
@@ -1001,7 +1006,7 @@
|
|
1001
1006
|
variants: function, method
|
1002
1007
|
dispatch:
|
1003
1008
|
SparseCPU, SparseCUDA: atan_sparse_
|
1004
|
-
SparseCsrCPU, SparseCsrCUDA: atan_sparse_csr_
|
1009
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: atan_sparse_csr_
|
1005
1010
|
tags: pointwise
|
1006
1011
|
|
1007
1012
|
- func: atan.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)
|
@@ -1012,7 +1017,7 @@
|
|
1012
1017
|
CPU, CUDA: atan_out
|
1013
1018
|
MPS: atan_out_mps
|
1014
1019
|
SparseCPU, SparseCUDA: atan_sparse_out
|
1015
|
-
SparseCsrCPU, SparseCsrCUDA: atan_sparse_csr_out
|
1020
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: atan_sparse_csr_out
|
1016
1021
|
tags: pointwise
|
1017
1022
|
|
1018
1023
|
# arctan, alias of atan
|
@@ -1423,7 +1428,7 @@
|
|
1423
1428
|
variants: function, method
|
1424
1429
|
dispatch:
|
1425
1430
|
SparseCPU, SparseCUDA: ceil_sparse
|
1426
|
-
SparseCsrCPU, SparseCsrCUDA: ceil_sparse_csr
|
1431
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: ceil_sparse_csr
|
1427
1432
|
tags: [core, pointwise]
|
1428
1433
|
|
1429
1434
|
- func: ceil_(Tensor(a!) self) -> Tensor(a!)
|
@@ -1432,7 +1437,7 @@
|
|
1432
1437
|
variants: function, method
|
1433
1438
|
dispatch:
|
1434
1439
|
SparseCPU, SparseCUDA: ceil_sparse_
|
1435
|
-
SparseCsrCPU, SparseCsrCUDA: ceil_sparse_csr_
|
1440
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: ceil_sparse_csr_
|
1436
1441
|
tags: pointwise
|
1437
1442
|
|
1438
1443
|
- func: ceil.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)
|
@@ -1443,7 +1448,7 @@
|
|
1443
1448
|
CPU, CUDA: ceil_out
|
1444
1449
|
MPS: ceil_out_mps
|
1445
1450
|
SparseCPU, SparseCUDA: ceil_sparse_out
|
1446
|
-
SparseCsrCPU, SparseCsrCUDA: ceil_sparse_csr_out
|
1451
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: ceil_sparse_csr_out
|
1447
1452
|
tags: pointwise
|
1448
1453
|
|
1449
1454
|
# alias for torch.linalg.multi_dot
|
@@ -1762,7 +1767,7 @@
|
|
1762
1767
|
MkldnnCPU: copy_mkldnn_
|
1763
1768
|
SparseCPU, SparseCUDA: copy_sparse_wrapper_
|
1764
1769
|
CompositeExplicitAutograd: copy_
|
1765
|
-
SparseCsrCPU, SparseCsrCUDA: copy_sparse_compressed_
|
1770
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: copy_sparse_compressed_
|
1766
1771
|
NestedTensorCPU, NestedTensorCUDA: copy_nested_
|
1767
1772
|
autogen: copy.out
|
1768
1773
|
|
@@ -2338,7 +2343,7 @@
|
|
2338
2343
|
|
2339
2344
|
- func: _embedding_bag_backward(Tensor grad, Tensor indices, Tensor offsets, Tensor offset2bag, Tensor bag_size, Tensor maximum_indices, SymInt num_weights, bool scale_grad_by_freq, int mode, bool sparse, Tensor? per_sample_weights, int padding_idx=-1) -> Tensor
|
2340
2345
|
dispatch:
|
2341
|
-
|
2346
|
+
CPU, CUDA: _embedding_bag_backward_symint
|
2342
2347
|
|
2343
2348
|
- func: _embedding_bag_sparse_backward(Tensor grad, Tensor indices, Tensor offsets, Tensor offset2bag, Tensor bag_size, SymInt num_weights, bool scale_grad_by_freq, int mode, Tensor? per_sample_weights, int padding_idx=-1) -> Tensor
|
2344
2349
|
dispatch:
|
@@ -2370,8 +2375,10 @@
|
|
2370
2375
|
MPS: empty_mps
|
2371
2376
|
Meta: empty_meta_symint
|
2372
2377
|
MkldnnCPU: empty_mkldnn
|
2373
|
-
SparseCPU, SparseCUDA
|
2374
|
-
|
2378
|
+
SparseCPU, SparseCUDA: empty_sparse
|
2379
|
+
SparseMeta: empty_sparse_symint
|
2380
|
+
SparseCsrCPU, SparseCsrCUDA: empty_sparse_compressed
|
2381
|
+
SparseCsrMeta: empty_sparse_compressed_symint
|
2375
2382
|
QuantizedCPU, QuantizedCUDA, QuantizedMeta: empty_unknown_quantized
|
2376
2383
|
tags: core
|
2377
2384
|
|
@@ -2446,7 +2453,7 @@
|
|
2446
2453
|
CUDA: resize_cuda_
|
2447
2454
|
MPS: resize_mps_
|
2448
2455
|
QuantizedCPU: quantized_resize_cpu_
|
2449
|
-
SparseCsrCPU, SparseCsrCUDA: resize_sparse_csr_
|
2456
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: resize_sparse_csr_
|
2450
2457
|
autogen: resize, resize.out
|
2451
2458
|
|
2452
2459
|
# This is a utility function to enable users to resize out tensor while registering kernels for out variants.
|
@@ -2497,7 +2504,7 @@
|
|
2497
2504
|
variants: function, method
|
2498
2505
|
dispatch:
|
2499
2506
|
SparseCPU, SparseCUDA: erf_sparse
|
2500
|
-
SparseCsrCPU, SparseCsrCUDA: erf_sparse_csr
|
2507
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: erf_sparse_csr
|
2501
2508
|
tags: [core, pointwise]
|
2502
2509
|
|
2503
2510
|
- func: erf_(Tensor(a!) self) -> Tensor(a!)
|
@@ -2506,7 +2513,7 @@
|
|
2506
2513
|
variants: function, method
|
2507
2514
|
dispatch:
|
2508
2515
|
SparseCPU, SparseCUDA: erf_sparse_
|
2509
|
-
SparseCsrCPU, SparseCsrCUDA: erf_sparse_csr_
|
2516
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: erf_sparse_csr_
|
2510
2517
|
tags: pointwise
|
2511
2518
|
|
2512
2519
|
- func: erf.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)
|
@@ -2517,7 +2524,7 @@
|
|
2517
2524
|
CPU, CUDA: erf_out
|
2518
2525
|
MPS: erf_out_mps
|
2519
2526
|
SparseCPU, SparseCUDA: erf_sparse_out
|
2520
|
-
SparseCsrCPU, SparseCsrCUDA: erf_sparse_csr_out
|
2527
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: erf_sparse_csr_out
|
2521
2528
|
tags: pointwise
|
2522
2529
|
|
2523
2530
|
- func: erfc(Tensor self) -> Tensor
|
@@ -2585,7 +2592,7 @@
|
|
2585
2592
|
variants: function, method
|
2586
2593
|
dispatch:
|
2587
2594
|
SparseCPU, SparseCUDA: expm1_sparse
|
2588
|
-
SparseCsrCPU, SparseCsrCUDA: expm1_sparse_csr
|
2595
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: expm1_sparse_csr
|
2589
2596
|
tags: [core, pointwise]
|
2590
2597
|
|
2591
2598
|
- func: expm1_(Tensor(a!) self) -> Tensor(a!)
|
@@ -2594,7 +2601,7 @@
|
|
2594
2601
|
variants: function, method
|
2595
2602
|
dispatch:
|
2596
2603
|
SparseCPU, SparseCUDA: expm1_sparse_
|
2597
|
-
SparseCsrCPU, SparseCsrCUDA: expm1_sparse_csr_
|
2604
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: expm1_sparse_csr_
|
2598
2605
|
tags: pointwise
|
2599
2606
|
|
2600
2607
|
- func: expm1.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)
|
@@ -2605,7 +2612,7 @@
|
|
2605
2612
|
CPU, CUDA: expm1_out
|
2606
2613
|
MPS: expm1_out_mps
|
2607
2614
|
SparseCPU, SparseCUDA: expm1_sparse_out
|
2608
|
-
SparseCsrCPU, SparseCsrCUDA: expm1_sparse_csr_out
|
2615
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: expm1_sparse_csr_out
|
2609
2616
|
tags: pointwise
|
2610
2617
|
|
2611
2618
|
- func: expand(Tensor(a) self, SymInt[] size, *, bool implicit=False) -> Tensor(a)
|
@@ -2683,7 +2690,7 @@
|
|
2683
2690
|
MPS: fill_scalar_mps
|
2684
2691
|
QuantizedCPU, QuantizedCUDA: fill_quantized_
|
2685
2692
|
Meta: fill_meta_
|
2686
|
-
SparseCsrCPU, SparseCsrCUDA: fill_sparse_csr_
|
2693
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: fill_sparse_csr_
|
2687
2694
|
NestedTensorCPU, NestedTensorCUDA: fill_nested_
|
2688
2695
|
autogen: fill.Scalar_out
|
2689
2696
|
|
@@ -2704,7 +2711,7 @@
|
|
2704
2711
|
variants: function, method
|
2705
2712
|
dispatch:
|
2706
2713
|
SparseCPU, SparseCUDA: floor_sparse
|
2707
|
-
SparseCsrCPU, SparseCsrCUDA: floor_sparse_csr
|
2714
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: floor_sparse_csr
|
2708
2715
|
tags: [core, pointwise]
|
2709
2716
|
|
2710
2717
|
- func: floor_(Tensor(a!) self) -> Tensor(a!)
|
@@ -2713,7 +2720,7 @@
|
|
2713
2720
|
variants: function, method
|
2714
2721
|
dispatch:
|
2715
2722
|
SparseCPU, SparseCUDA: floor_sparse_
|
2716
|
-
SparseCsrCPU, SparseCsrCUDA: floor_sparse_csr_
|
2723
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: floor_sparse_csr_
|
2717
2724
|
tags: pointwise
|
2718
2725
|
|
2719
2726
|
- func: floor.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)
|
@@ -2724,7 +2731,7 @@
|
|
2724
2731
|
CPU, CUDA: floor_out
|
2725
2732
|
MPS: floor_out_mps
|
2726
2733
|
SparseCPU, SparseCUDA: floor_sparse_out
|
2727
|
-
SparseCsrCPU, SparseCsrCUDA: floor_sparse_csr_out
|
2734
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: floor_sparse_csr_out
|
2728
2735
|
tags: pointwise
|
2729
2736
|
|
2730
2737
|
- func: floor_divide(Tensor self, Tensor other) -> Tensor
|
@@ -2769,7 +2776,7 @@
|
|
2769
2776
|
variants: function, method
|
2770
2777
|
dispatch:
|
2771
2778
|
SparseCPU, SparseCUDA: frac_sparse
|
2772
|
-
SparseCsrCPU, SparseCsrCUDA: frac_sparse_csr
|
2779
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: frac_sparse_csr
|
2773
2780
|
tags: pointwise
|
2774
2781
|
|
2775
2782
|
- func: frac_(Tensor(a!) self) -> Tensor(a!)
|
@@ -2778,7 +2785,7 @@
|
|
2778
2785
|
variants: function, method
|
2779
2786
|
dispatch:
|
2780
2787
|
SparseCPU, SparseCUDA: frac_sparse_
|
2781
|
-
SparseCsrCPU, SparseCsrCUDA: frac_sparse_csr_
|
2788
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: frac_sparse_csr_
|
2782
2789
|
tags: pointwise
|
2783
2790
|
|
2784
2791
|
- func: frac.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)
|
@@ -2789,7 +2796,7 @@
|
|
2789
2796
|
CPU, CUDA: frac_out
|
2790
2797
|
MPS: frac_out_mps
|
2791
2798
|
SparseCPU, SparseCUDA: frac_sparse_out
|
2792
|
-
SparseCsrCPU, SparseCsrCUDA: frac_sparse_csr_out
|
2799
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: frac_sparse_csr_out
|
2793
2800
|
tags: pointwise
|
2794
2801
|
|
2795
2802
|
- func: full.names(int[] size, Scalar fill_value, *, Dimname[]? names, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
|
@@ -3061,6 +3068,18 @@
|
|
3061
3068
|
dispatch:
|
3062
3069
|
CompositeExplicitAutograd: _unsafe_index
|
3063
3070
|
|
3071
|
+
# Used by inductor to generate masked loads
|
3072
|
+
# Note that we don't support boolean indexing, to avoid dynamic output shapes
|
3073
|
+
- func: _unsafe_masked_index(Tensor self, Tensor mask, Tensor?[] indices, Scalar fill) -> Tensor
|
3074
|
+
variants: function
|
3075
|
+
dispatch:
|
3076
|
+
CompositeExplicitAutograd: _unsafe_masked_index
|
3077
|
+
|
3078
|
+
- func: _unsafe_masked_index_put_accumulate(Tensor self, Tensor mask, Tensor?[] indices, Tensor values) -> Tensor
|
3079
|
+
variants: function
|
3080
|
+
dispatch:
|
3081
|
+
CompositeExplicitAutograd: _unsafe_masked_index_put_accumulate
|
3082
|
+
|
3064
3083
|
- func: index_copy.out(Tensor self, int dim, Tensor index, Tensor source, *, Tensor(a!) out) -> Tensor(a!)
|
3065
3084
|
structured: True
|
3066
3085
|
variants: function
|
@@ -3161,7 +3180,7 @@
|
|
3161
3180
|
dispatch:
|
3162
3181
|
CPU, CUDA, MPS: isnan
|
3163
3182
|
SparseCPU, SparseCUDA: isnan_sparse
|
3164
|
-
SparseCsrCPU, SparseCsrCUDA: isnan_sparse_csr
|
3183
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: isnan_sparse_csr
|
3165
3184
|
autogen: isnan.out
|
3166
3185
|
tags: [core, pointwise]
|
3167
3186
|
|
@@ -3381,6 +3400,10 @@
|
|
3381
3400
|
|
3382
3401
|
- func: fbgemm_pack_gemm_matrix_fp16(Tensor input) -> Tensor
|
3383
3402
|
|
3403
|
+
- func: _wrapped_linear_prepack(Tensor weight, Tensor weight_scale, Tensor weight_zero_point, Tensor bias) -> Tensor
|
3404
|
+
|
3405
|
+
- func: _wrapped_quantized_linear_prepacked(Tensor input, Tensor input_scale, Tensor input_zero_point, Tensor packed_weight, Tensor output_scale, Tensor output_zero_point, int out_channel) -> Tensor
|
3406
|
+
|
3384
3407
|
- func: fbgemm_linear_fp16_weight_fp32_activation(Tensor input, Tensor packed_weight, Tensor bias) -> Tensor
|
3385
3408
|
|
3386
3409
|
- func: fbgemm_linear_fp16_weight(Tensor input, Tensor packed_weight, Tensor bias) -> Tensor
|
@@ -3487,7 +3510,7 @@
|
|
3487
3510
|
variants: function, method
|
3488
3511
|
dispatch:
|
3489
3512
|
SparseCPU, SparseCUDA: log1p_sparse
|
3490
|
-
SparseCsrCPU, SparseCsrCUDA: log1p_sparse_csr
|
3513
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: log1p_sparse_csr
|
3491
3514
|
tags: [core, pointwise]
|
3492
3515
|
|
3493
3516
|
- func: log1p_(Tensor(a!) self) -> Tensor(a!)
|
@@ -3496,7 +3519,7 @@
|
|
3496
3519
|
variants: function, method
|
3497
3520
|
dispatch:
|
3498
3521
|
SparseCPU, SparseCUDA: log1p_sparse_
|
3499
|
-
SparseCsrCPU, SparseCsrCUDA: log1p_sparse_csr_
|
3522
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: log1p_sparse_csr_
|
3500
3523
|
tags: pointwise
|
3501
3524
|
|
3502
3525
|
- func: log1p.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)
|
@@ -3507,7 +3530,7 @@
|
|
3507
3530
|
CPU, CUDA: log1p_out
|
3508
3531
|
MPS: log1p_out_mps
|
3509
3532
|
SparseCPU, SparseCUDA: log1p_sparse_out
|
3510
|
-
SparseCsrCPU, SparseCsrCUDA: log1p_sparse_csr_out
|
3533
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: log1p_sparse_csr_out
|
3511
3534
|
tags: pointwise
|
3512
3535
|
|
3513
3536
|
- func: log2(Tensor self) -> Tensor
|
@@ -3899,11 +3922,10 @@
|
|
3899
3922
|
tags: core
|
3900
3923
|
|
3901
3924
|
# For normal naming convention this should be `mean.out`. However since we already have `mean.out` we have to rename this.
|
3902
|
-
|
3903
|
-
|
3904
|
-
|
3905
|
-
|
3906
|
-
# CompositeExplicitAutograd: mean_dtype_out
|
3925
|
+
- func: mean.dtype_out(Tensor self, *, ScalarType? dtype=None, Tensor(a!) out) -> Tensor(a!)
|
3926
|
+
device_check: NoCheck # TensorIterator
|
3927
|
+
dispatch:
|
3928
|
+
CompositeExplicitAutograd: mean_dtype_out
|
3907
3929
|
|
3908
3930
|
- func: mean.dim(Tensor self, int[1]? dim, bool keepdim=False, *, ScalarType? dtype=None) -> Tensor
|
3909
3931
|
structured_delegate: mean.out
|
@@ -4095,7 +4117,7 @@
|
|
4095
4117
|
variants: function, method
|
4096
4118
|
dispatch:
|
4097
4119
|
SparseCPU, SparseCUDA: _sparse_mm
|
4098
|
-
SparseCsrCPU, SparseCsrCUDA: _sparse_csr_mm
|
4120
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: _sparse_csr_mm
|
4099
4121
|
tags: core
|
4100
4122
|
|
4101
4123
|
- func: mm.out(Tensor self, Tensor mat2, *, Tensor(a!) out) -> Tensor(a!)
|
@@ -4105,7 +4127,7 @@
|
|
4105
4127
|
CUDA: mm_out_cuda
|
4106
4128
|
MPS: mm_out_mps
|
4107
4129
|
SparseCPU, SparseCUDA: _sparse_mm_out
|
4108
|
-
SparseCsrCPU, SparseCsrCUDA: _sparse_csr_mm_out
|
4130
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: _sparse_csr_mm_out
|
4109
4131
|
|
4110
4132
|
- func: _int_mm(Tensor self, Tensor mat2) -> Tensor
|
4111
4133
|
dispatch:
|
@@ -4121,6 +4143,7 @@
|
|
4121
4143
|
dispatch:
|
4122
4144
|
CPU: _convert_weight_to_int4pack_cpu
|
4123
4145
|
CUDA: _convert_weight_to_int4pack_cuda
|
4146
|
+
MPS: _convert_weight_to_int4pack_mps
|
4124
4147
|
|
4125
4148
|
- func: _weight_int4pack_mm(Tensor self, Tensor mat2, int qGroupSize, Tensor qScaleAndZeros) -> Tensor
|
4126
4149
|
dispatch:
|
@@ -4165,7 +4188,7 @@
|
|
4165
4188
|
variants: function, method
|
4166
4189
|
dispatch:
|
4167
4190
|
SparseCPU, SparseCUDA: mul_sparse
|
4168
|
-
SparseCsrCPU, SparseCsrCUDA: mul_sparse_csr
|
4191
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: mul_sparse_csr
|
4169
4192
|
MkldnnCPU: mkldnn_mul
|
4170
4193
|
ZeroTensor: mul_zerotensor
|
4171
4194
|
NestedTensorCPU, NestedTensorCUDA: NestedTensor_mul_Tensor
|
@@ -4177,7 +4200,7 @@
|
|
4177
4200
|
variants: method
|
4178
4201
|
dispatch:
|
4179
4202
|
SparseCPU, SparseCUDA: mul_sparse_
|
4180
|
-
SparseCsrCPU, SparseCsrCUDA: mul_sparse_csr_
|
4203
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: mul_sparse_csr_
|
4181
4204
|
MkldnnCPU: mkldnn_mul_
|
4182
4205
|
NestedTensorCPU, NestedTensorCUDA: NestedTensor_mul__Tensor
|
4183
4206
|
tags: pointwise
|
@@ -4191,7 +4214,7 @@
|
|
4191
4214
|
MPS: mul_out_mps
|
4192
4215
|
SparseCPU: mul_out_sparse_cpu
|
4193
4216
|
SparseCUDA: mul_out_sparse_cuda
|
4194
|
-
SparseCsrCPU, SparseCsrCUDA: mul_out_sparse_csr
|
4217
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: mul_out_sparse_csr
|
4195
4218
|
MkldnnCPU: mkldnn_mul_out
|
4196
4219
|
tags: pointwise
|
4197
4220
|
# For C++ only, until we have conversion from C++ numbers to Tensor
|
@@ -4201,7 +4224,7 @@
|
|
4201
4224
|
variants: function, method
|
4202
4225
|
dispatch:
|
4203
4226
|
CompositeExplicitAutograd: mul
|
4204
|
-
SparseCsrCPU, SparseCsrCUDA: mul_scalar_sparse_csr
|
4227
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: mul_scalar_sparse_csr
|
4205
4228
|
NestedTensorCPU, NestedTensorCUDA: NestedTensor_mul_Scalar
|
4206
4229
|
tags: [core, pointwise]
|
4207
4230
|
|
@@ -4210,7 +4233,7 @@
|
|
4210
4233
|
variants: method
|
4211
4234
|
dispatch:
|
4212
4235
|
CompositeExplicitAutograd: mul_
|
4213
|
-
SparseCsrCPU, SparseCsrCUDA: mul__scalar_sparse_csr
|
4236
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: mul__scalar_sparse_csr
|
4214
4237
|
NestedTensorCPU, NestedTensorCUDA: NestedTensor_mul__Scalar
|
4215
4238
|
autogen: mul.Scalar_out
|
4216
4239
|
tags: pointwise
|
@@ -4530,9 +4553,11 @@
|
|
4530
4553
|
- func: is_pinned(Tensor self, Device? device=None) -> bool
|
4531
4554
|
variants: method
|
4532
4555
|
dispatch:
|
4533
|
-
|
4534
|
-
|
4535
|
-
CompositeExplicitAutograd:
|
4556
|
+
# the NestedTensor keys are necessary because NestedTensor has been removed
|
4557
|
+
# from the CompositeExplicitAutograd keyset see Note [NestedTensor Not Included in Backend Keys]
|
4558
|
+
CompositeExplicitAutograd, NestedTensorCPU: is_pinned
|
4559
|
+
SparseCsrCPU: is_pinned_sparse_compressed
|
4560
|
+
SparseCPU: is_pinned_sparse_coo
|
4536
4561
|
|
4537
4562
|
# TODO: add a copy kwarg that guarantees that the tensor is put into fresh
|
4538
4563
|
# pinned memory
|
@@ -4542,9 +4567,10 @@
|
|
4542
4567
|
# Unlike pin_memory, this is guaranteed to give a new non-aliasing tensor
|
4543
4568
|
- func: _pin_memory(Tensor self, Device? device=None) -> Tensor
|
4544
4569
|
dispatch:
|
4545
|
-
|
4546
|
-
|
4547
|
-
|
4570
|
+
CompositeExplicitAutograd: _pin_memory
|
4571
|
+
NestedTensorCPU: _pin_memory_nested
|
4572
|
+
SparseCPU: _pin_memory_sparse_coo
|
4573
|
+
SparseCsrCPU: _pin_memory_sparse_compressed
|
4548
4574
|
autogen: _pin_memory.out
|
4549
4575
|
|
4550
4576
|
- func: pinverse(Tensor self, float rcond=1e-15) -> Tensor
|
@@ -4558,27 +4584,27 @@
|
|
4558
4584
|
dispatch:
|
4559
4585
|
CompositeExplicitAutograd: rad2deg
|
4560
4586
|
SparseCPU, SparseCUDA: rad2deg_sparse
|
4561
|
-
SparseCsrCPU, SparseCsrCUDA: rad2deg_sparse_csr
|
4587
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: rad2deg_sparse_csr
|
4562
4588
|
|
4563
4589
|
- func: rad2deg_(Tensor(a!) self) -> Tensor(a!)
|
4564
4590
|
variants: function, method
|
4565
4591
|
dispatch:
|
4566
4592
|
CompositeExplicitAutograd: rad2deg_
|
4567
4593
|
SparseCPU, SparseCUDA: rad2deg_sparse_
|
4568
|
-
SparseCsrCPU, SparseCsrCUDA: rad2deg_sparse_csr_
|
4594
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: rad2deg_sparse_csr_
|
4569
4595
|
|
4570
4596
|
- func: rad2deg.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)
|
4571
4597
|
dispatch:
|
4572
4598
|
CompositeExplicitAutograd: rad2deg_out
|
4573
4599
|
SparseCPU, SparseCUDA: rad2deg_sparse_out
|
4574
|
-
SparseCsrCPU, SparseCsrCUDA: rad2deg_sparse_csr_out
|
4600
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: rad2deg_sparse_csr_out
|
4575
4601
|
|
4576
4602
|
- func: deg2rad(Tensor self) -> Tensor
|
4577
4603
|
variants: function, method
|
4578
4604
|
dispatch:
|
4579
4605
|
CompositeExplicitAutograd: deg2rad
|
4580
4606
|
SparseCPU, SparseCUDA: deg2rad_sparse
|
4581
|
-
SparseCsrCPU, SparseCsrCUDA: deg2rad_sparse_csr
|
4607
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: deg2rad_sparse_csr
|
4582
4608
|
tags: pointwise
|
4583
4609
|
|
4584
4610
|
- func: deg2rad_(Tensor(a!) self) -> Tensor(a!)
|
@@ -4586,14 +4612,14 @@
|
|
4586
4612
|
dispatch:
|
4587
4613
|
CompositeExplicitAutograd: deg2rad_
|
4588
4614
|
SparseCPU, SparseCUDA: deg2rad_sparse_
|
4589
|
-
SparseCsrCPU, SparseCsrCUDA: deg2rad_sparse_csr_
|
4615
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: deg2rad_sparse_csr_
|
4590
4616
|
tags: pointwise
|
4591
4617
|
|
4592
4618
|
- func: deg2rad.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)
|
4593
4619
|
dispatch:
|
4594
4620
|
CompositeExplicitAutograd: deg2rad_out
|
4595
4621
|
SparseCPU, SparseCUDA: deg2rad_sparse_out
|
4596
|
-
SparseCsrCPU, SparseCsrCUDA: deg2rad_sparse_csr_out
|
4622
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: deg2rad_sparse_csr_out
|
4597
4623
|
tags: pointwise
|
4598
4624
|
|
4599
4625
|
- func: scalar_tensor(Scalar s, *, ScalarType? dtype=None, Layout? layout=None, Device? device=None, bool? pin_memory=None) -> Tensor
|
@@ -4811,7 +4837,7 @@
|
|
4811
4837
|
variants: function, method
|
4812
4838
|
dispatch:
|
4813
4839
|
SparseCPU, SparseCUDA: neg_sparse
|
4814
|
-
SparseCsrCPU, SparseCsrCUDA: neg_sparse_csr
|
4840
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: neg_sparse_csr
|
4815
4841
|
NestedTensorCPU, NestedTensorCUDA: NestedTensor_neg
|
4816
4842
|
tags: [core, pointwise]
|
4817
4843
|
|
@@ -4821,7 +4847,7 @@
|
|
4821
4847
|
variants: function, method
|
4822
4848
|
dispatch:
|
4823
4849
|
SparseCPU, SparseCUDA: neg_sparse_
|
4824
|
-
SparseCsrCPU, SparseCsrCUDA: neg_sparse_csr_
|
4850
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: neg_sparse_csr_
|
4825
4851
|
NestedTensorCPU, NestedTensorCUDA: NestedTensor_neg_
|
4826
4852
|
tags: pointwise
|
4827
4853
|
|
@@ -4833,7 +4859,7 @@
|
|
4833
4859
|
CPU, CUDA: neg_out
|
4834
4860
|
MPS: neg_out_mps
|
4835
4861
|
SparseCPU, SparseCUDA: neg_out_sparse
|
4836
|
-
SparseCsrCPU, SparseCsrCUDA: neg_sparse_csr_out
|
4862
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: neg_sparse_csr_out
|
4837
4863
|
tags: pointwise
|
4838
4864
|
# Alias for neg
|
4839
4865
|
|
@@ -4917,7 +4943,7 @@
|
|
4917
4943
|
variants: function, method
|
4918
4944
|
dispatch:
|
4919
4945
|
SparseCPU, SparseCUDA: round_sparse
|
4920
|
-
SparseCsrCPU, SparseCsrCUDA: round_sparse_csr
|
4946
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: round_sparse_csr
|
4921
4947
|
tags: [core, pointwise]
|
4922
4948
|
|
4923
4949
|
- func: round_(Tensor(a!) self) -> Tensor(a!)
|
@@ -4926,7 +4952,7 @@
|
|
4926
4952
|
variants: function, method
|
4927
4953
|
dispatch:
|
4928
4954
|
SparseCPU, SparseCUDA: round_sparse_
|
4929
|
-
SparseCsrCPU, SparseCsrCUDA: round_sparse_csr_
|
4955
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: round_sparse_csr_
|
4930
4956
|
tags: pointwise
|
4931
4957
|
|
4932
4958
|
- func: round.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)
|
@@ -4938,7 +4964,7 @@
|
|
4938
4964
|
CUDA: round_out
|
4939
4965
|
MPS: round_out_mps
|
4940
4966
|
SparseCPU, SparseCUDA: round_sparse_out
|
4941
|
-
SparseCsrCPU, SparseCsrCUDA: round_sparse_csr_out
|
4967
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: round_sparse_csr_out
|
4942
4968
|
tags: pointwise
|
4943
4969
|
|
4944
4970
|
- func: round.decimals(Tensor self, *, int decimals) -> Tensor
|
@@ -4981,7 +5007,7 @@
|
|
4981
5007
|
QuantizedCUDA: relu_quantized_cuda
|
4982
5008
|
NestedTensorCPU, NestedTensorCUDA: NestedTensor_relu
|
4983
5009
|
SparseCPU, SparseCUDA: relu_sparse
|
4984
|
-
SparseCsrCPU, SparseCsrCUDA: relu_sparse_csr
|
5010
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: relu_sparse_csr
|
4985
5011
|
tags: [core, pointwise]
|
4986
5012
|
|
4987
5013
|
- func: relu_(Tensor(a!) self) -> Tensor(a!)
|
@@ -4995,7 +5021,7 @@
|
|
4995
5021
|
QuantizedCUDA: relu_quantized_cuda_
|
4996
5022
|
NestedTensorCPU, NestedTensorCUDA: NestedTensor_relu_
|
4997
5023
|
SparseCPU, SparseCUDA: relu_sparse_
|
4998
|
-
SparseCsrCPU, SparseCsrCUDA: relu_sparse_csr_
|
5024
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: relu_sparse_csr_
|
4999
5025
|
autogen: relu.out
|
5000
5026
|
tags: pointwise
|
5001
5027
|
|
@@ -5128,7 +5154,7 @@
|
|
5128
5154
|
device_guard: False
|
5129
5155
|
dispatch:
|
5130
5156
|
CompositeExplicitAutograd: select_symint
|
5131
|
-
SparseCsrCPU, SparseCsrCUDA: select_sparse_csr
|
5157
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: select_sparse_csr
|
5132
5158
|
NestedTensorCPU, NestedTensorCUDA: select_nested
|
5133
5159
|
tags: core
|
5134
5160
|
|
@@ -5277,7 +5303,7 @@
|
|
5277
5303
|
structured_delegate: sin.out
|
5278
5304
|
variants: function, method
|
5279
5305
|
dispatch:
|
5280
|
-
SparseCsrCPU, SparseCsrCUDA: sin_sparse_csr
|
5306
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: sin_sparse_csr
|
5281
5307
|
SparseCPU, SparseCUDA: sin_sparse
|
5282
5308
|
NestedTensorCPU, NestedTensorCUDA: sin_nested
|
5283
5309
|
tags: [core, pointwise]
|
@@ -5287,7 +5313,7 @@
|
|
5287
5313
|
structured_delegate: sin.out
|
5288
5314
|
variants: function, method
|
5289
5315
|
dispatch:
|
5290
|
-
SparseCsrCPU, SparseCsrCUDA: sin_sparse_csr_
|
5316
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: sin_sparse_csr_
|
5291
5317
|
SparseCPU, SparseCUDA: sin_sparse_
|
5292
5318
|
tags: pointwise
|
5293
5319
|
|
@@ -5298,7 +5324,7 @@
|
|
5298
5324
|
dispatch:
|
5299
5325
|
CPU, CUDA: sin_out
|
5300
5326
|
MPS: sin_out_mps
|
5301
|
-
SparseCsrCPU, SparseCsrCUDA: sin_sparse_csr_out
|
5327
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: sin_sparse_csr_out
|
5302
5328
|
SparseCPU, SparseCUDA: sin_sparse_out
|
5303
5329
|
tags: pointwise
|
5304
5330
|
|
@@ -5325,7 +5351,7 @@
|
|
5325
5351
|
variants: function, method
|
5326
5352
|
dispatch:
|
5327
5353
|
SparseCPU, SparseCUDA: sinh_sparse
|
5328
|
-
SparseCsrCPU, SparseCsrCUDA: sinh_sparse_csr
|
5354
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: sinh_sparse_csr
|
5329
5355
|
tags: [core, pointwise]
|
5330
5356
|
|
5331
5357
|
- func: sinh_(Tensor(a!) self) -> Tensor(a!)
|
@@ -5334,7 +5360,7 @@
|
|
5334
5360
|
variants: function, method
|
5335
5361
|
dispatch:
|
5336
5362
|
SparseCPU, SparseCUDA: sinh_sparse_
|
5337
|
-
SparseCsrCPU, SparseCsrCUDA: sinh_sparse_csr_
|
5363
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: sinh_sparse_csr_
|
5338
5364
|
tags: pointwise
|
5339
5365
|
|
5340
5366
|
- func: sinh.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)
|
@@ -5345,7 +5371,7 @@
|
|
5345
5371
|
CPU, CUDA: sinh_out
|
5346
5372
|
MPS: sinh_out_mps
|
5347
5373
|
SparseCPU, SparseCUDA: sinh_sparse_out
|
5348
|
-
SparseCsrCPU, SparseCsrCUDA: sinh_sparse_csr_out
|
5374
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: sinh_sparse_csr_out
|
5349
5375
|
|
5350
5376
|
# Returns a copy of this `Variable` that is detached from its autograd graph.
|
5351
5377
|
# This method is OK to call if the `Variable` is a view.
|
@@ -5732,7 +5758,7 @@
|
|
5732
5758
|
dispatch:
|
5733
5759
|
NestedTensorCPU: NestedTensor_sum_dim_CPU
|
5734
5760
|
SparseCPU, SparseCUDA: sum_sparse_coo
|
5735
|
-
SparseCsrCPU, SparseCsrCUDA: sum_sparse_compressed
|
5761
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: sum_sparse_compressed
|
5736
5762
|
tags: core
|
5737
5763
|
|
5738
5764
|
- func: sum.dim_DimnameList(Tensor self, Dimname[1] dim, bool keepdim=False, *, ScalarType? dtype=None) -> Tensor
|
@@ -5778,7 +5804,7 @@
|
|
5778
5804
|
variants: function, method
|
5779
5805
|
dispatch:
|
5780
5806
|
SparseCPU, SparseCUDA: sqrt_sparse
|
5781
|
-
SparseCsrCPU, SparseCsrCUDA: sqrt_sparse_csr
|
5807
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: sqrt_sparse_csr
|
5782
5808
|
tags: [core, pointwise]
|
5783
5809
|
|
5784
5810
|
- func: sqrt_(Tensor(a!) self) -> Tensor(a!)
|
@@ -5787,7 +5813,7 @@
|
|
5787
5813
|
variants: function, method
|
5788
5814
|
dispatch:
|
5789
5815
|
SparseCPU, SparseCUDA: sqrt_sparse_
|
5790
|
-
SparseCsrCPU, SparseCsrCUDA: sqrt_sparse_csr_
|
5816
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: sqrt_sparse_csr_
|
5791
5817
|
tags: pointwise
|
5792
5818
|
|
5793
5819
|
- func: sqrt.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)
|
@@ -5798,7 +5824,7 @@
|
|
5798
5824
|
CPU, CUDA: sqrt_out
|
5799
5825
|
MPS: sqrt_out_mps
|
5800
5826
|
SparseCPU, SparseCUDA: sqrt_sparse_out
|
5801
|
-
SparseCsrCPU, SparseCsrCUDA: sqrt_sparse_csr_out
|
5827
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: sqrt_sparse_csr_out
|
5802
5828
|
tags: pointwise
|
5803
5829
|
|
5804
5830
|
- func: square(Tensor self) -> Tensor
|
@@ -5936,7 +5962,7 @@
|
|
5936
5962
|
variants: function, method
|
5937
5963
|
dispatch:
|
5938
5964
|
SparseCPU, SparseCUDA: tan_sparse
|
5939
|
-
SparseCsrCPU, SparseCsrCUDA: tan_sparse_csr
|
5965
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: tan_sparse_csr
|
5940
5966
|
tags: [core, pointwise]
|
5941
5967
|
|
5942
5968
|
- func: tan_(Tensor(a!) self) -> Tensor(a!)
|
@@ -5945,7 +5971,7 @@
|
|
5945
5971
|
variants: function, method
|
5946
5972
|
dispatch:
|
5947
5973
|
SparseCPU, SparseCUDA: tan_sparse_
|
5948
|
-
SparseCsrCPU, SparseCsrCUDA: tan_sparse_csr_
|
5974
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: tan_sparse_csr_
|
5949
5975
|
tags: pointwise
|
5950
5976
|
|
5951
5977
|
- func: tan.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)
|
@@ -5956,7 +5982,7 @@
|
|
5956
5982
|
CPU, CUDA: tan_out
|
5957
5983
|
MPS: tan_out_mps
|
5958
5984
|
SparseCPU, SparseCUDA: tan_sparse_out
|
5959
|
-
SparseCsrCPU, SparseCsrCUDA: tan_sparse_csr_out
|
5985
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: tan_sparse_csr_out
|
5960
5986
|
tags: pointwise
|
5961
5987
|
|
5962
5988
|
- func: tanh(Tensor self) -> Tensor
|
@@ -5967,7 +5993,7 @@
|
|
5967
5993
|
QuantizedCPU: tanh_quantized_cpu
|
5968
5994
|
MkldnnCPU: mkldnn_tanh
|
5969
5995
|
SparseCPU, SparseCUDA: tanh_sparse
|
5970
|
-
SparseCsrCPU, SparseCsrCUDA: tanh_sparse_csr
|
5996
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: tanh_sparse_csr
|
5971
5997
|
NestedTensorCPU, NestedTensorCUDA: NestedTensor_tanh
|
5972
5998
|
tags: [core, pointwise]
|
5973
5999
|
|
@@ -5978,7 +6004,7 @@
|
|
5978
6004
|
dispatch:
|
5979
6005
|
MkldnnCPU: mkldnn_tanh_
|
5980
6006
|
SparseCPU, SparseCUDA: tanh_sparse_
|
5981
|
-
SparseCsrCPU, SparseCsrCUDA: tanh_sparse_csr_
|
6007
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: tanh_sparse_csr_
|
5982
6008
|
NestedTensorCPU, NestedTensorCUDA: NestedTensor_tanh_
|
5983
6009
|
tags: pointwise
|
5984
6010
|
|
@@ -5990,7 +6016,7 @@
|
|
5990
6016
|
CPU, CUDA: tanh_out
|
5991
6017
|
MPS: tanh_out_mps
|
5992
6018
|
SparseCPU, SparseCUDA: tanh_sparse_out
|
5993
|
-
SparseCsrCPU, SparseCsrCUDA: tanh_sparse_csr_out
|
6019
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: tanh_sparse_csr_out
|
5994
6020
|
tags: pointwise
|
5995
6021
|
|
5996
6022
|
- func: tensordot(Tensor self, Tensor other, int[] dims_self, int[] dims_other) -> Tensor
|
@@ -6027,7 +6053,7 @@
|
|
6027
6053
|
CPU, CUDA: threshold_backward_out
|
6028
6054
|
MPS: threshold_backward_out_mps
|
6029
6055
|
SparseCPU, SparseCUDA: threshold_backward_sparse_out
|
6030
|
-
SparseCsrCPU, SparseCsrCUDA: threshold_backward_sparse_compressed_out
|
6056
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: threshold_backward_sparse_compressed_out
|
6031
6057
|
|
6032
6058
|
- func: threshold_backward(Tensor grad_output, Tensor self, Scalar threshold) -> Tensor
|
6033
6059
|
variants: function
|
@@ -6035,7 +6061,7 @@
|
|
6035
6061
|
dispatch:
|
6036
6062
|
MkldnnCPU: mkldnn_relu_backward
|
6037
6063
|
SparseCPU, SparseCUDA: threshold_backward_sparse
|
6038
|
-
SparseCsrCPU, SparseCsrCUDA: threshold_backward_sparse_compressed
|
6064
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: threshold_backward_sparse_compressed
|
6039
6065
|
NestedTensorCPU, NestedTensorCUDA: threshold_backwards_nested
|
6040
6066
|
tags: pointwise
|
6041
6067
|
|
@@ -6185,12 +6211,12 @@
|
|
6185
6211
|
CompositeExplicitAutogradNonFunctional: _nested_view_from_buffer_copy
|
6186
6212
|
autogen: _nested_view_from_buffer_copy.out
|
6187
6213
|
|
6188
|
-
- func: _nested_view_from_jagged(Tensor(a) self, Tensor offsets, Tensor dummy, Tensor? lengths=None, int ragged_idx=1) -> Tensor(a)
|
6214
|
+
- func: _nested_view_from_jagged(Tensor(a) self, Tensor offsets, Tensor dummy, Tensor? lengths=None, int ragged_idx=1, Tensor? min_seqlen=None, Tensor? max_seqlen=None) -> Tensor(a)
|
6189
6215
|
variants: function
|
6190
6216
|
device_check: NoCheck
|
6191
6217
|
dispatch: {}
|
6192
6218
|
|
6193
|
-
- func: _nested_view_from_jagged_copy(Tensor self, Tensor offsets, Tensor dummy, Tensor? lengths=None, int ragged_idx=1) -> Tensor
|
6219
|
+
- func: _nested_view_from_jagged_copy(Tensor self, Tensor offsets, Tensor dummy, Tensor? lengths=None, int ragged_idx=1, Tensor? min_seqlen=None, Tensor? max_seqlen=None) -> Tensor
|
6194
6220
|
variants: function
|
6195
6221
|
device_check: NoCheck
|
6196
6222
|
tags: view_copy
|
@@ -6227,6 +6253,16 @@
|
|
6227
6253
|
device_check: NoCheck
|
6228
6254
|
dispatch: {}
|
6229
6255
|
|
6256
|
+
- func: _nested_get_min_seqlen(Tensor self) -> Tensor
|
6257
|
+
variants: function
|
6258
|
+
device_check: NoCheck
|
6259
|
+
dispatch: {}
|
6260
|
+
|
6261
|
+
- func: _nested_get_max_seqlen(Tensor self) -> Tensor
|
6262
|
+
variants: function
|
6263
|
+
device_check: NoCheck
|
6264
|
+
dispatch: {}
|
6265
|
+
|
6230
6266
|
- func: _nested_get_jagged_dummy(Tensor any) -> Tensor
|
6231
6267
|
category_override: dummy
|
6232
6268
|
dispatch: {}
|
@@ -6251,7 +6287,7 @@
|
|
6251
6287
|
variants: function, method
|
6252
6288
|
dispatch:
|
6253
6289
|
SparseCPU, SparseCUDA: trunc_sparse
|
6254
|
-
SparseCsrCPU, SparseCsrCUDA: trunc_sparse_csr
|
6290
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: trunc_sparse_csr
|
6255
6291
|
tags: [core, pointwise]
|
6256
6292
|
|
6257
6293
|
- func: trunc_(Tensor(a!) self) -> Tensor(a!)
|
@@ -6260,7 +6296,7 @@
|
|
6260
6296
|
variants: function, method
|
6261
6297
|
dispatch:
|
6262
6298
|
SparseCPU, SparseCUDA: trunc_sparse_
|
6263
|
-
SparseCsrCPU, SparseCsrCUDA: trunc_sparse_csr_
|
6299
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: trunc_sparse_csr_
|
6264
6300
|
tags: pointwise
|
6265
6301
|
|
6266
6302
|
- func: trunc.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)
|
@@ -6271,7 +6307,7 @@
|
|
6271
6307
|
CPU, CUDA: trunc_out
|
6272
6308
|
MPS: trunc_out_mps
|
6273
6309
|
SparseCPU, SparseCUDA: trunc_sparse_out
|
6274
|
-
SparseCsrCPU, SparseCsrCUDA: trunc_sparse_csr_out
|
6310
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: trunc_sparse_csr_out
|
6275
6311
|
tags: pointwise
|
6276
6312
|
# Alias for trunc
|
6277
6313
|
|
@@ -6443,6 +6479,7 @@
|
|
6443
6479
|
variants: function, method
|
6444
6480
|
dispatch:
|
6445
6481
|
CPU, CUDA, MPS: where
|
6482
|
+
NestedTensorCPU, NestedTensorCUDA: NestedTensor_where
|
6446
6483
|
tags: [core, pointwise]
|
6447
6484
|
|
6448
6485
|
- func: where.self_out(Tensor condition, Tensor self, Tensor other, *, Tensor(a!) out) -> Tensor(a!)
|
@@ -6780,7 +6817,7 @@
|
|
6780
6817
|
dispatch:
|
6781
6818
|
CompositeExplicitAutograd: clone
|
6782
6819
|
SparseCPU, SparseCUDA: clone_sparse
|
6783
|
-
SparseCsrCPU, SparseCsrCUDA: clone_sparse_compressed
|
6820
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: clone_sparse_compressed
|
6784
6821
|
MkldnnCPU: mkldnn_clone
|
6785
6822
|
QuantizedCPU, QuantizedCUDA: quantized_clone
|
6786
6823
|
NestedTensorCPU, NestedTensorCUDA: clone_nested
|
@@ -6804,7 +6841,7 @@
|
|
6804
6841
|
variants: function, method
|
6805
6842
|
dispatch:
|
6806
6843
|
SparseCPU, SparseCUDA: resize_as_sparse_
|
6807
|
-
SparseCsrCPU, SparseCsrCUDA: resize_as_sparse_compressed_
|
6844
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: resize_as_sparse_compressed_
|
6808
6845
|
autogen: resize_as_sparse, resize_as_sparse.out
|
6809
6846
|
|
6810
6847
|
- func: zero_(Tensor(a!) self) -> Tensor(a!)
|
@@ -6962,7 +6999,7 @@
|
|
6962
6999
|
dispatch:
|
6963
7000
|
SparseCPU: addmm_sparse_dense_cpu
|
6964
7001
|
SparseCUDA: addmm_sparse_dense_cuda
|
6965
|
-
SparseCsrCPU, SparseCsrCUDA: addmm_sparse_compressed_dense
|
7002
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: addmm_sparse_compressed_dense
|
6966
7003
|
tags: core
|
6967
7004
|
|
6968
7005
|
- func: addmm_(Tensor(a!) self, Tensor mat1, Tensor mat2, *, Scalar beta=1, Scalar alpha=1) -> Tensor(a!)
|
@@ -6984,12 +7021,12 @@
|
|
6984
7021
|
structured_delegate: _addmm_activation.out
|
6985
7022
|
variants: function, method
|
6986
7023
|
|
6987
|
-
- func: _scaled_mm(Tensor self, Tensor mat2,
|
7024
|
+
- func: _scaled_mm(Tensor self, Tensor mat2, Tensor scale_a, Tensor scale_b, Tensor? bias=None, Tensor? scale_result=None, ScalarType? out_dtype=None, bool use_fast_accum=False) -> Tensor
|
6988
7025
|
variants: function
|
6989
7026
|
dispatch:
|
6990
7027
|
CUDA: _scaled_mm_cuda
|
6991
7028
|
|
6992
|
-
- func: _scaled_mm.out(Tensor self, Tensor mat2,
|
7029
|
+
- func: _scaled_mm.out(Tensor self, Tensor mat2, Tensor scale_a, Tensor scale_b, Tensor? bias=None, Tensor? scale_result=None, ScalarType? out_dtype=None, bool use_fast_accum=False, *, Tensor(a!) out) -> Tensor(a!)
|
6993
7030
|
variants: function
|
6994
7031
|
dispatch:
|
6995
7032
|
CUDA: _scaled_mm_out_cuda
|
@@ -7184,7 +7221,7 @@
|
|
7184
7221
|
variants: method
|
7185
7222
|
dispatch:
|
7186
7223
|
SparseCPU, SparseCUDA: sparse_mask
|
7187
|
-
SparseCsrCPU, SparseCsrCUDA: sparse_mask_sparse_compressed
|
7224
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: sparse_mask_sparse_compressed
|
7188
7225
|
autogen: sparse_mask.out
|
7189
7226
|
|
7190
7227
|
- func: _sparse_mask_projection(Tensor self, Tensor mask, bool accumulate_matches=False) -> Tensor
|
@@ -7204,7 +7241,7 @@
|
|
7204
7241
|
variants: method
|
7205
7242
|
dispatch:
|
7206
7243
|
SparseCPU, SparseCUDA: sparse_to_dense
|
7207
|
-
SparseCsrCPU, SparseCsrCUDA: sparse_compressed_to_dense
|
7244
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: sparse_compressed_to_dense
|
7208
7245
|
MkldnnCPU: mkldnn_to_dense
|
7209
7246
|
autogen: _to_dense.out
|
7210
7247
|
|
@@ -7385,7 +7422,7 @@
|
|
7385
7422
|
dispatch:
|
7386
7423
|
CPU, CUDA: dense_to_sparse
|
7387
7424
|
SparseCPU, SparseCUDA: sparse_coo_to_sparse
|
7388
|
-
SparseCsrCPU, SparseCsrCUDA: sparse_compressed_to_sparse
|
7425
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: sparse_compressed_to_sparse
|
7389
7426
|
autogen: _to_sparse.sparse_dim_out
|
7390
7427
|
|
7391
7428
|
- func: to_sparse(Tensor self, *, Layout? layout=None, int[2]? blocksize=None, int? dense_dim=None) -> Tensor
|
@@ -7397,7 +7434,7 @@
|
|
7397
7434
|
dispatch:
|
7398
7435
|
CPU, CUDA: dense_to_sparse
|
7399
7436
|
SparseCPU, SparseCUDA: sparse_coo_to_sparse
|
7400
|
-
SparseCsrCPU, SparseCsrCUDA: sparse_compressed_to_sparse
|
7437
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: sparse_compressed_to_sparse
|
7401
7438
|
autogen: _to_sparse.out
|
7402
7439
|
|
7403
7440
|
- func: to_sparse_csr(Tensor self, int? dense_dim=None) -> Tensor
|
@@ -7409,7 +7446,7 @@
|
|
7409
7446
|
dispatch:
|
7410
7447
|
CPU, CUDA: dense_to_sparse_csr
|
7411
7448
|
SparseCPU, SparseCUDA: coo_to_sparse_csr
|
7412
|
-
SparseCsrCPU, SparseCsrCUDA: sparse_compressed_to_sparse_csr
|
7449
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: sparse_compressed_to_sparse_csr
|
7413
7450
|
autogen: _to_sparse_csr.out
|
7414
7451
|
|
7415
7452
|
- func: to_sparse_csc(Tensor self, int? dense_dim=None) -> Tensor
|
@@ -7421,7 +7458,7 @@
|
|
7421
7458
|
dispatch:
|
7422
7459
|
CPU, CUDA: dense_to_sparse_csc
|
7423
7460
|
SparseCPU, SparseCUDA: coo_to_sparse_csc
|
7424
|
-
SparseCsrCPU, SparseCsrCUDA: sparse_compressed_to_sparse_csc
|
7461
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: sparse_compressed_to_sparse_csc
|
7425
7462
|
autogen: _to_sparse_csc.out
|
7426
7463
|
|
7427
7464
|
- func: to_sparse_bsr(Tensor self, int[2] blocksize, int? dense_dim=None) -> Tensor
|
@@ -7433,7 +7470,7 @@
|
|
7433
7470
|
dispatch:
|
7434
7471
|
CPU, CUDA: dense_to_sparse_bsr
|
7435
7472
|
SparseCPU, SparseCUDA: coo_to_sparse_bsr
|
7436
|
-
SparseCsrCPU, SparseCsrCUDA: sparse_compressed_to_sparse_bsr
|
7473
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: sparse_compressed_to_sparse_bsr
|
7437
7474
|
autogen: _to_sparse_bsr.out
|
7438
7475
|
|
7439
7476
|
- func: to_sparse_bsc(Tensor self, int[2] blocksize, int? dense_dim=None) -> Tensor
|
@@ -7445,7 +7482,7 @@
|
|
7445
7482
|
dispatch:
|
7446
7483
|
CPU, CUDA: dense_to_sparse_bsc
|
7447
7484
|
SparseCPU, SparseCUDA: coo_to_sparse_bsc
|
7448
|
-
SparseCsrCPU, SparseCsrCUDA: sparse_compressed_to_sparse_bsc
|
7485
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: sparse_compressed_to_sparse_bsc
|
7449
7486
|
autogen: _to_sparse_bsc.out
|
7450
7487
|
|
7451
7488
|
- func: _to_sparse_semi_structured(Tensor dense) -> (Tensor, Tensor)
|
@@ -8431,21 +8468,21 @@
|
|
8431
8468
|
device_check: NoCheck # TensorIterator
|
8432
8469
|
variants: method, function
|
8433
8470
|
dispatch:
|
8434
|
-
CPU, CUDA: __lshift__
|
8471
|
+
CPU, CUDA, MPS: __lshift__
|
8435
8472
|
tags: pointwise
|
8436
8473
|
|
8437
8474
|
- func: __lshift__.Tensor(Tensor self, Tensor other) -> Tensor
|
8438
8475
|
device_check: NoCheck # TensorIterator
|
8439
8476
|
variants: method, function
|
8440
8477
|
dispatch:
|
8441
|
-
CPU, CUDA: __lshift__
|
8478
|
+
CPU, CUDA, MPS: __lshift__
|
8442
8479
|
tags: pointwise
|
8443
8480
|
|
8444
8481
|
- func: __ilshift__.Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)
|
8445
8482
|
device_check: NoCheck # TensorIterator
|
8446
8483
|
variants: method
|
8447
8484
|
dispatch:
|
8448
|
-
CPU, CUDA: __ilshift__
|
8485
|
+
CPU, CUDA, MPS: __ilshift__
|
8449
8486
|
autogen: __lshift__.Scalar_out
|
8450
8487
|
tags: pointwise
|
8451
8488
|
|
@@ -8453,7 +8490,7 @@
|
|
8453
8490
|
device_check: NoCheck # TensorIterator
|
8454
8491
|
variants: method
|
8455
8492
|
dispatch:
|
8456
|
-
CPU, CUDA: __ilshift__
|
8493
|
+
CPU, CUDA, MPS: __ilshift__
|
8457
8494
|
autogen: __lshift__.Tensor_out
|
8458
8495
|
tags: pointwise
|
8459
8496
|
|
@@ -8474,7 +8511,7 @@
|
|
8474
8511
|
structured: True
|
8475
8512
|
structured_inherits: TensorIteratorBase
|
8476
8513
|
dispatch:
|
8477
|
-
CPU, CUDA: bitwise_left_shift_out
|
8514
|
+
CPU, CUDA, MPS: bitwise_left_shift_out
|
8478
8515
|
tags: pointwise
|
8479
8516
|
|
8480
8517
|
- func: bitwise_left_shift.Tensor_Scalar(Tensor self, Scalar other) -> Tensor
|
@@ -8510,28 +8547,28 @@
|
|
8510
8547
|
device_check: NoCheck # TensorIterator
|
8511
8548
|
variants: method, function
|
8512
8549
|
dispatch:
|
8513
|
-
CPU, CUDA: __rshift__
|
8550
|
+
CPU, CUDA, MPS: __rshift__
|
8514
8551
|
tags: pointwise
|
8515
8552
|
|
8516
8553
|
- func: __rshift__.Tensor(Tensor self, Tensor other) -> Tensor
|
8517
8554
|
device_check: NoCheck # TensorIterator
|
8518
8555
|
variants: method, function
|
8519
8556
|
dispatch:
|
8520
|
-
CPU, CUDA: __rshift__
|
8557
|
+
CPU, CUDA, MPS: __rshift__
|
8521
8558
|
tags: pointwise
|
8522
8559
|
|
8523
8560
|
- func: __irshift__.Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)
|
8524
8561
|
device_check: NoCheck # TensorIterator
|
8525
8562
|
variants: method
|
8526
8563
|
dispatch:
|
8527
|
-
CPU, CUDA: __irshift__
|
8564
|
+
CPU, CUDA, MPS: __irshift__
|
8528
8565
|
autogen: __rshift__.Scalar_out
|
8529
8566
|
|
8530
8567
|
- func: __irshift__.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)
|
8531
8568
|
device_check: NoCheck # TensorIterator
|
8532
8569
|
variants: method
|
8533
8570
|
dispatch:
|
8534
|
-
CPU, CUDA: __irshift__
|
8571
|
+
CPU, CUDA, MPS: __irshift__
|
8535
8572
|
autogen: __rshift__.Tensor_out
|
8536
8573
|
|
8537
8574
|
- func: bitwise_right_shift.Tensor(Tensor self, Tensor other) -> Tensor
|
@@ -8551,7 +8588,7 @@
|
|
8551
8588
|
structured: True
|
8552
8589
|
structured_inherits: TensorIteratorBase
|
8553
8590
|
dispatch:
|
8554
|
-
CPU, CUDA: bitwise_right_shift_out
|
8591
|
+
CPU, CUDA, MPS: bitwise_right_shift_out
|
8555
8592
|
tags: pointwise
|
8556
8593
|
|
8557
8594
|
- func: bitwise_right_shift.Tensor_Scalar(Tensor self, Scalar other) -> Tensor
|
@@ -8858,6 +8895,7 @@
|
|
8858
8895
|
variants: method, function
|
8859
8896
|
dispatch:
|
8860
8897
|
QuantizedCPU: eq_quantized_cpu
|
8898
|
+
NestedTensorCPU, NestedTensorCUDA: eq_tensor_nested
|
8861
8899
|
tags: [core, pointwise]
|
8862
8900
|
|
8863
8901
|
- func: ge.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!)
|
@@ -9502,7 +9540,7 @@
|
|
9502
9540
|
variants: method, function
|
9503
9541
|
dispatch:
|
9504
9542
|
SparseCPU, SparseCUDA: erfinv_sparse
|
9505
|
-
SparseCsrCPU, SparseCsrCUDA: erfinv_sparse_csr
|
9543
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: erfinv_sparse_csr
|
9506
9544
|
tags: pointwise
|
9507
9545
|
|
9508
9546
|
- func: erfinv_(Tensor(a!) self) -> Tensor(a!)
|
@@ -9511,7 +9549,7 @@
|
|
9511
9549
|
variants: method
|
9512
9550
|
dispatch:
|
9513
9551
|
SparseCPU, SparseCUDA: erfinv_sparse_
|
9514
|
-
SparseCsrCPU, SparseCsrCUDA: erfinv_sparse_csr_
|
9552
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: erfinv_sparse_csr_
|
9515
9553
|
tags: pointwise
|
9516
9554
|
|
9517
9555
|
- func: erfinv.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)
|
@@ -9522,7 +9560,7 @@
|
|
9522
9560
|
CPU, CUDA: erfinv_out
|
9523
9561
|
MPS: erfinv_out_mps
|
9524
9562
|
SparseCPU, SparseCUDA: erfinv_sparse_out
|
9525
|
-
SparseCsrCPU, SparseCsrCUDA: erfinv_sparse_csr_out
|
9563
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: erfinv_sparse_csr_out
|
9526
9564
|
tags: pointwise
|
9527
9565
|
|
9528
9566
|
- func: i0(Tensor self) -> Tensor
|
@@ -9548,7 +9586,7 @@
|
|
9548
9586
|
variants: function, method
|
9549
9587
|
dispatch:
|
9550
9588
|
SparseCPU, SparseCUDA: sign_sparse
|
9551
|
-
SparseCsrCPU, SparseCsrCUDA: sign_sparse_csr
|
9589
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: sign_sparse_csr
|
9552
9590
|
tags: [core, pointwise]
|
9553
9591
|
|
9554
9592
|
- func: sign_(Tensor(a!) self) -> Tensor(a!)
|
@@ -9557,7 +9595,7 @@
|
|
9557
9595
|
variants: method
|
9558
9596
|
dispatch:
|
9559
9597
|
SparseCPU, SparseCUDA: sign_sparse_
|
9560
|
-
SparseCsrCPU, SparseCsrCUDA: sign_sparse_csr_
|
9598
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: sign_sparse_csr_
|
9561
9599
|
tags: pointwise
|
9562
9600
|
|
9563
9601
|
- func: sign.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)
|
@@ -9568,7 +9606,7 @@
|
|
9568
9606
|
CPU, CUDA: sign_out
|
9569
9607
|
MPS: sign_out_mps
|
9570
9608
|
SparseCPU, SparseCUDA: sign_sparse_out
|
9571
|
-
SparseCsrCPU, SparseCsrCUDA: sign_sparse_csr_out
|
9609
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: sign_sparse_csr_out
|
9572
9610
|
tags: pointwise
|
9573
9611
|
|
9574
9612
|
- func: signbit(Tensor self) -> Tensor
|
@@ -9576,7 +9614,7 @@
|
|
9576
9614
|
structured_delegate: signbit.out
|
9577
9615
|
dispatch:
|
9578
9616
|
SparseCPU, SparseCUDA: signbit_sparse
|
9579
|
-
SparseCsrCPU, SparseCsrCUDA: signbit_sparse_csr
|
9617
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: signbit_sparse_csr
|
9580
9618
|
tags: pointwise
|
9581
9619
|
|
9582
9620
|
- func: signbit.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)
|
@@ -9587,7 +9625,7 @@
|
|
9587
9625
|
CUDA: signbit_out
|
9588
9626
|
MPS: signbit_out_mps
|
9589
9627
|
SparseCPU, SparseCUDA: signbit_sparse_out
|
9590
|
-
SparseCsrCPU, SparseCsrCUDA: signbit_sparse_csr_out
|
9628
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: signbit_sparse_csr_out
|
9591
9629
|
tags: pointwise
|
9592
9630
|
|
9593
9631
|
- func: dist(Tensor self, Tensor other, Scalar p=2) -> Tensor
|
@@ -10038,9 +10076,10 @@
|
|
10038
10076
|
- func: argsort.stable(Tensor self, *, bool stable, int dim=-1, bool descending=False) -> Tensor
|
10039
10077
|
device_check: NoCheck # TensorIterator
|
10040
10078
|
variants: method, function
|
10041
|
-
|
10042
|
-
|
10043
|
-
|
10079
|
+
|
10080
|
+
- func: argsort.stable_out(Tensor self, *, bool stable, int dim=-1, bool descending=False, Tensor(a!) out) -> Tensor(a!)
|
10081
|
+
device_check: NoCheck # TensorIterator
|
10082
|
+
variants: function
|
10044
10083
|
|
10045
10084
|
- func: argsort.dimname(Tensor self, Dimname dim, bool descending=False) -> Tensor
|
10046
10085
|
variants: method, function
|
@@ -10220,7 +10259,7 @@
|
|
10220
10259
|
CPU, CUDA: normal_
|
10221
10260
|
MPS: normal_mps_
|
10222
10261
|
Meta: normal_meta_
|
10223
|
-
SparseCsrCPU, SparseCsrCUDA: normal_sparse_csr_
|
10262
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: normal_sparse_csr_
|
10224
10263
|
NestedTensorCPU, NestedTensorCUDA: normal_nested_
|
10225
10264
|
autogen: normal.out
|
10226
10265
|
|
@@ -13024,7 +13063,7 @@
|
|
13024
13063
|
CompositeExplicitAutograd: isinf
|
13025
13064
|
SparseCPU, SparseCUDA: isinf_sparse
|
13026
13065
|
SparseMeta: isinf_sparse_meta
|
13027
|
-
SparseCsrCPU, SparseCsrCUDA: isinf_sparse_csr
|
13066
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: isinf_sparse_csr
|
13028
13067
|
autogen: isinf.out
|
13029
13068
|
tags: [core, pointwise]
|
13030
13069
|
|
@@ -13038,7 +13077,7 @@
|
|
13038
13077
|
structured_delegate: isposinf.out
|
13039
13078
|
dispatch:
|
13040
13079
|
SparseCPU, SparseCUDA: isposinf_sparse
|
13041
|
-
SparseCsrCPU, SparseCsrCUDA: isposinf_sparse_csr
|
13080
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: isposinf_sparse_csr
|
13042
13081
|
tags: pointwise
|
13043
13082
|
|
13044
13083
|
- func: isposinf.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)
|
@@ -13047,7 +13086,7 @@
|
|
13047
13086
|
dispatch:
|
13048
13087
|
CPU, CUDA: isposinf_out
|
13049
13088
|
SparseCPU, SparseCUDA: isposinf_sparse_out
|
13050
|
-
SparseCsrCPU, SparseCsrCUDA: isposinf_sparse_csr_out
|
13089
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: isposinf_sparse_csr_out
|
13051
13090
|
tags: pointwise
|
13052
13091
|
|
13053
13092
|
- func: isneginf(Tensor self) -> Tensor
|
@@ -13055,7 +13094,7 @@
|
|
13055
13094
|
structured_delegate: isneginf.out
|
13056
13095
|
dispatch:
|
13057
13096
|
SparseCPU, SparseCUDA: isneginf_sparse
|
13058
|
-
SparseCsrCPU, SparseCsrCUDA: isneginf_sparse_csr
|
13097
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: isneginf_sparse_csr
|
13059
13098
|
tags: pointwise
|
13060
13099
|
|
13061
13100
|
- func: isneginf.out(Tensor self, *, Tensor(a!) out) -> Tensor(a!)
|
@@ -13064,7 +13103,7 @@
|
|
13064
13103
|
dispatch:
|
13065
13104
|
CPU, CUDA: isneginf_out
|
13066
13105
|
SparseCPU, SparseCUDA: isneginf_sparse_out
|
13067
|
-
SparseCsrCPU, SparseCsrCUDA: isneginf_sparse_csr_out
|
13106
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: isneginf_sparse_csr_out
|
13068
13107
|
tags: pointwise
|
13069
13108
|
|
13070
13109
|
# NOTE [_add_batch_dim and _remove_batch_dim]
|
@@ -13787,10 +13826,16 @@
|
|
13787
13826
|
- func: linalg_lu_factor(Tensor A, *, bool pivot=True) -> (Tensor LU, Tensor pivots)
|
13788
13827
|
python_module: linalg
|
13789
13828
|
variants: function
|
13829
|
+
dispatch:
|
13830
|
+
CompositeImplicitAutograd: linalg_lu_factor
|
13831
|
+
MPS: linalg_lu_factor_mps
|
13790
13832
|
|
13791
13833
|
- func: linalg_lu_factor.out(Tensor A, *, bool pivot=True, Tensor(a!) LU, Tensor(b!) pivots) -> (Tensor(a!) LU, Tensor(b!) pivots)
|
13792
13834
|
python_module: linalg
|
13793
13835
|
variants: function
|
13836
|
+
dispatch:
|
13837
|
+
CompositeImplicitAutograd: linalg_lu_factor_out
|
13838
|
+
MPS: linalg_lu_factor_out_mps
|
13794
13839
|
|
13795
13840
|
- func: linalg_lu_factor_ex(Tensor A, *, bool pivot=True, bool check_errors=False) -> (Tensor LU, Tensor pivots, Tensor info)
|
13796
13841
|
python_module: linalg
|
@@ -14176,6 +14221,11 @@
|
|
14176
14221
|
- func: linalg_solve(Tensor A, Tensor B, *, bool left=True) -> Tensor
|
14177
14222
|
python_module: linalg
|
14178
14223
|
|
14224
|
+
- func: _spsolve(Tensor A, Tensor B, *, bool left=True) -> Tensor
|
14225
|
+
python_module: sparse
|
14226
|
+
dispatch:
|
14227
|
+
SparseCsrCUDA: _sparse_csr_linear_solve
|
14228
|
+
|
14179
14229
|
- func: linalg_solve.out(Tensor A, Tensor B, *, bool left=True, Tensor(a!) out) -> Tensor(a!)
|
14180
14230
|
python_module: linalg
|
14181
14231
|
|
@@ -14352,7 +14402,7 @@
|
|
14352
14402
|
CPU, CUDA: _segment_reduce_backward_kernel
|
14353
14403
|
autogen: _segment_reduce_backward.out
|
14354
14404
|
|
14355
|
-
- func: pad_sequence(Tensor[] sequences, bool batch_first=False, float padding_value=0.0) -> Tensor
|
14405
|
+
- func: pad_sequence(Tensor[] sequences, bool batch_first=False, float padding_value=0.0, str padding_side="right") -> Tensor
|
14356
14406
|
python_module: nn
|
14357
14407
|
variants: function
|
14358
14408
|
|
@@ -14458,7 +14508,7 @@
|
|
14458
14508
|
variants: function
|
14459
14509
|
dispatch:
|
14460
14510
|
CompositeExplicitAutogradNonFunctional: select_copy_symint
|
14461
|
-
SparseCsrCPU, SparseCsrCUDA: select_copy_sparse_csr
|
14511
|
+
SparseCsrCPU, SparseCsrCUDA, SparseCsrMeta: select_copy_sparse_csr
|
14462
14512
|
tags: view_copy
|
14463
14513
|
autogen: select_copy.int_out
|
14464
14514
|
|
@@ -14648,11 +14698,13 @@
|
|
14648
14698
|
variants: function
|
14649
14699
|
dispatch:
|
14650
14700
|
CUDA: _fbgemm_jagged_to_padded_dense_forward
|
14701
|
+
CPU: _jagged_to_padded_dense_forward_cpu
|
14651
14702
|
|
14652
14703
|
- func: _padded_dense_to_jagged_forward(Tensor dense, Tensor[] offsets, SymInt? total_L=None) -> Tensor
|
14653
14704
|
variants: function
|
14654
14705
|
dispatch:
|
14655
14706
|
CUDA: _fbgemm_dense_to_jagged_forward_symint
|
14707
|
+
CPU: _padded_dense_to_jagged_forward_cpu
|
14656
14708
|
|
14657
14709
|
- func: _nested_tensor_softmax_with_shape(Tensor self, Tensor query) -> Tensor
|
14658
14710
|
dispatch:
|
@@ -14660,6 +14712,11 @@
|
|
14660
14712
|
NestedTensorCUDA: NestedTensor_softmax_dropout_cuda
|
14661
14713
|
tags: nondeterministic_seeded
|
14662
14714
|
|
14715
|
+
- func: _safe_softmax(Tensor self, int dim, ScalarType? dtype=None) -> Tensor
|
14716
|
+
dispatch:
|
14717
|
+
CompositeExplicitAutograd: _safe_softmax
|
14718
|
+
NestedTensorCPU, NestedTensorCUDA: _safe_softmax
|
14719
|
+
|
14663
14720
|
# Apparently, putting "forward" in the name will cause Python bindings to be skipped, so "fwd" it is.
|
14664
14721
|
- func: _transformer_encoder_layer_fwd(Tensor src, int embed_dim, int num_heads, Tensor qkv_weight, Tensor qkv_bias, Tensor proj_weight, Tensor proj_bias, bool use_gelu, bool norm_first, float eps, Tensor norm_weight_1, Tensor norm_bias_1, Tensor norm_weight_2, Tensor norm_bias_2, Tensor ffn_weight_1, Tensor ffn_bias_1, Tensor ffn_weight_2, Tensor ffn_bias_2, Tensor? mask=None, int? mask_type=None) -> Tensor
|
14665
14722
|
variants: function
|
@@ -14674,24 +14731,29 @@
|
|
14674
14731
|
CUDA, NestedTensorCUDA: native_multi_head_attention_cuda
|
14675
14732
|
autogen: _native_multi_head_attention.out
|
14676
14733
|
|
14677
|
-
- func: scaled_dot_product_attention(Tensor query, Tensor key, Tensor value, Tensor? attn_mask=None, float dropout_p=0.0, bool is_causal=False, *, float? scale=None) -> Tensor
|
14734
|
+
- func: scaled_dot_product_attention(Tensor query, Tensor key, Tensor value, Tensor? attn_mask=None, float dropout_p=0.0, bool is_causal=False, *, float? scale=None, bool enable_gqa=False) -> Tensor
|
14678
14735
|
python_module: nn
|
14679
14736
|
variants: function
|
14680
14737
|
autogen: scaled_dot_product_attention.out
|
14681
14738
|
tags: nondeterministic_seeded
|
14682
14739
|
|
14683
14740
|
# This aten function is kept so that we can test the choice function from Python
|
14684
|
-
- func: _fused_sdp_choice(Tensor query, Tensor key, Tensor value, Tensor? attn_mask=None, float dropout_p=0.0, bool is_causal=False, *, float? scale=None) -> int
|
14741
|
+
- func: _fused_sdp_choice(Tensor query, Tensor key, Tensor value, Tensor? attn_mask=None, float dropout_p=0.0, bool is_causal=False, *, float? scale=None, bool enable_gqa=False) -> int
|
14685
14742
|
dispatch:
|
14686
14743
|
Meta: _fused_sdp_choice_meta
|
14687
14744
|
CPU, NestedTensorCPU: _fused_sdp_choice_cpp
|
14688
14745
|
CUDA, NestedTensorCUDA: _fused_sdp_choice_cuda
|
14689
14746
|
tags: nondeterministic_seeded
|
14690
14747
|
|
14691
|
-
- func: _scaled_dot_product_attention_math(Tensor query, Tensor key, Tensor value, Tensor? attn_mask=None, float dropout_p=0.0, bool is_causal=False, Tensor? dropout_mask=None, *, float? scale=None) -> (Tensor, Tensor)
|
14748
|
+
- func: _scaled_dot_product_attention_math(Tensor query, Tensor key, Tensor value, Tensor? attn_mask=None, float dropout_p=0.0, bool is_causal=False, Tensor? dropout_mask=None, *, float? scale=None, bool enable_gqa=False) -> (Tensor, Tensor)
|
14692
14749
|
variants: function
|
14693
14750
|
tags: nondeterministic_seeded
|
14694
14751
|
|
14752
|
+
- func: _scaled_dot_product_attention_math_for_mps(Tensor query, Tensor key, Tensor value, Tensor? attn_mask=None, float dropout_p=0.0, bool is_causal=False, Tensor? dropout_mask=None, *, float? scale=None) -> (Tensor, Tensor)
|
14753
|
+
dispatch:
|
14754
|
+
MPS: _scaled_dot_product_attention_math_mps
|
14755
|
+
tags: nondeterministic_seeded
|
14756
|
+
|
14695
14757
|
- func: _scaled_dot_product_flash_attention(Tensor query, Tensor key, Tensor value, float dropout_p=0.0, bool is_causal=False, bool return_debug_mask=False, *, float? scale=None) -> (Tensor output, Tensor logsumexp, Tensor cum_seq_q, Tensor cum_seq_k, SymInt max_q, SymInt max_k, Tensor philox_seed, Tensor philox_offset, Tensor debug_attn_mask)
|
14696
14758
|
dispatch:
|
14697
14759
|
CUDA: _scaled_dot_product_flash_attention_cuda
|
@@ -14703,6 +14765,11 @@
|
|
14703
14765
|
CPU: _scaled_dot_product_flash_attention_cpu
|
14704
14766
|
tags: nondeterministic_seeded
|
14705
14767
|
|
14768
|
+
- func: _scaled_dot_product_fused_attention_overrideable(Tensor query, Tensor key, Tensor value, Tensor? attn_bias=None, float dropout_p=0.0, bool is_causal=False, bool return_debug_mask=False, *, float? scale=None) -> (Tensor output, Tensor logsumexp, Tensor cum_seq_q, Tensor cum_seq_k, SymInt max_q, SymInt max_k, Tensor philox_seed, Tensor philox_offset, Tensor debug_attn_mask)
|
14769
|
+
dispatch:
|
14770
|
+
CompositeExplicitAutograd: _scaled_dot_product_fused_attention_overrideable
|
14771
|
+
tags: nondeterministic_seeded
|
14772
|
+
|
14706
14773
|
- func: _scaled_dot_product_flash_attention_backward(Tensor grad_out, Tensor query, Tensor key, Tensor value, Tensor out, Tensor logsumexp, Tensor cum_seq_q, Tensor cum_seq_k, SymInt max_q, SymInt max_k, float dropout_p, bool is_causal, Tensor philox_seed, Tensor philox_offset, *, float? scale=None) -> (Tensor grad_query, Tensor grad_key, Tensor grad_value)
|
14707
14774
|
device_check: NoCheck
|
14708
14775
|
variants: function
|
@@ -14716,6 +14783,12 @@
|
|
14716
14783
|
dispatch:
|
14717
14784
|
CPU: _scaled_dot_product_flash_attention_cpu_backward
|
14718
14785
|
|
14786
|
+
- func: _scaled_dot_product_fused_attention_overrideable_backward(Tensor grad_out, Tensor query, Tensor key, Tensor value, Tensor attn_bias, bool[4] grad_input_mask, Tensor out, Tensor logsumexp, Tensor cum_seq_q, Tensor cum_seq_k, SymInt max_q, SymInt max_k, float dropout_p, bool is_causal, Tensor philox_seed, Tensor philox_offset, *, float? scale=None) -> (Tensor grad_query, Tensor grad_key, Tensor grad_value, Tensor grad_attn_bias)
|
14787
|
+
device_check: NoCheck
|
14788
|
+
variants: function
|
14789
|
+
dispatch:
|
14790
|
+
CompositeExplicitAutograd: _scaled_dot_product_fused_attention_overrideable_backward
|
14791
|
+
|
14719
14792
|
- func: _scaled_dot_product_efficient_attention(Tensor query, Tensor key, Tensor value, Tensor? attn_bias, bool compute_log_sumexp, float dropout_p=0.0, bool is_causal=False, *, float? scale=None) -> (Tensor output, Tensor log_sumexp, Tensor philox_seed, Tensor philox_offset)
|
14720
14793
|
dispatch:
|
14721
14794
|
CUDA: _scaled_dot_product_efficient_attention_cuda
|
@@ -14728,12 +14801,12 @@
|
|
14728
14801
|
CUDA: _scaled_dot_product_efficient_attention_backward_cuda
|
14729
14802
|
tags: nondeterministic_seeded
|
14730
14803
|
|
14731
|
-
- func: _scaled_dot_product_cudnn_attention(Tensor query, Tensor key, Tensor value, float dropout_p=0.0, bool is_causal=False, bool return_debug_mask=False, *, float? scale=None) -> (Tensor output, Tensor logsumexp, Tensor cum_seq_q, Tensor cum_seq_k, SymInt max_q, SymInt max_k, Tensor philox_seed, Tensor philox_offset, Tensor debug_attn_mask)
|
14804
|
+
- func: _scaled_dot_product_cudnn_attention(Tensor query, Tensor key, Tensor value, Tensor? attn_bias, bool compute_log_sumexp, float dropout_p=0.0, bool is_causal=False, bool return_debug_mask=False, *, float? scale=None) -> (Tensor output, Tensor logsumexp, Tensor cum_seq_q, Tensor cum_seq_k, SymInt max_q, SymInt max_k, Tensor philox_seed, Tensor philox_offset, Tensor debug_attn_mask)
|
14732
14805
|
dispatch:
|
14733
14806
|
CUDA: _scaled_dot_product_cudnn_attention_cuda
|
14734
14807
|
tags: nondeterministic_seeded
|
14735
14808
|
|
14736
|
-
- func: _scaled_dot_product_cudnn_attention_backward(Tensor grad_out, Tensor query, Tensor key, Tensor value, Tensor out, Tensor logsumexp, Tensor cum_seq_q, Tensor cum_seq_k, SymInt max_q, SymInt max_k, float dropout_p, bool is_causal,
|
14809
|
+
- func: _scaled_dot_product_cudnn_attention_backward(Tensor grad_out, Tensor query, Tensor key, Tensor value, Tensor out, Tensor logsumexp, Tensor philox_seed, Tensor philox_offset, Tensor attn_bias, Tensor cum_seq_q, Tensor cum_seq_k, SymInt max_q, SymInt max_k, float dropout_p, bool is_causal, *, float? scale=None) -> (Tensor, Tensor, Tensor)
|
14737
14810
|
dispatch:
|
14738
14811
|
CUDA: _scaled_dot_product_cudnn_attention_backward_cuda
|
14739
14812
|
tags: nondeterministic_seeded
|
@@ -15563,6 +15636,7 @@
|
|
15563
15636
|
dispatch:
|
15564
15637
|
CPU: _fused_adam_kernel_cpu_
|
15565
15638
|
CUDA: _fused_adam_kernel_cuda_
|
15639
|
+
MPS: _fused_adam_kernel_mps_
|
15566
15640
|
autogen: _fused_adam, _fused_adam.out
|
15567
15641
|
|
15568
15642
|
- func: _fused_adam_.tensor_lr(Tensor(a!)[] self, Tensor(b!)[] grads, Tensor(c!)[] exp_avgs, Tensor(d!)[] exp_avg_sqs, Tensor(e!)[] max_exp_avg_sqs, Tensor[] state_steps, *, Tensor lr, float beta1, float beta2, float weight_decay, float eps, bool amsgrad, bool maximize, Tensor? grad_scale=None, Tensor? found_inf=None) -> ()
|
@@ -15573,6 +15647,7 @@
|
|
15573
15647
|
dispatch:
|
15574
15648
|
CPU: _fused_adam_kernel_cpu_
|
15575
15649
|
CUDA: _fused_adam_kernel_cuda_
|
15650
|
+
MPS: _fused_adam_kernel_mps_
|
15576
15651
|
autogen: _fused_adam.tensor_lr, _fused_adam.tensor_lr_out
|
15577
15652
|
|
15578
15653
|
- func: _fused_adamw_(Tensor(a!)[] self, Tensor(b!)[] grads, Tensor(c!)[] exp_avgs, Tensor(d!)[] exp_avg_sqs, Tensor(e!)[] max_exp_avg_sqs, Tensor[] state_steps, *, float lr, float beta1, float beta2, float weight_decay, float eps, bool amsgrad, bool maximize, Tensor? grad_scale=None, Tensor? found_inf=None) -> ()
|
@@ -15581,6 +15656,7 @@
|
|
15581
15656
|
dispatch:
|
15582
15657
|
CPU: _fused_adamw_kernel_cpu_
|
15583
15658
|
CUDA: _fused_adamw_kernel_cuda_
|
15659
|
+
MPS: _fused_adamw_kernel_mps_
|
15584
15660
|
autogen: _fused_adamw, _fused_adamw.out
|
15585
15661
|
|
15586
15662
|
- func: _fused_adamw_.tensor_lr(Tensor(a!)[] self, Tensor(b!)[] grads, Tensor(c!)[] exp_avgs, Tensor(d!)[] exp_avg_sqs, Tensor(e!)[] max_exp_avg_sqs, Tensor[] state_steps, *, Tensor lr, float beta1, float beta2, float weight_decay, float eps, bool amsgrad, bool maximize, Tensor? grad_scale=None, Tensor? found_inf=None) -> ()
|
@@ -15591,6 +15667,7 @@
|
|
15591
15667
|
dispatch:
|
15592
15668
|
CPU: _fused_adamw_kernel_cpu_
|
15593
15669
|
CUDA: _fused_adamw_kernel_cuda_
|
15670
|
+
MPS: _fused_adamw_kernel_mps_
|
15594
15671
|
autogen: _fused_adamw.tensor_lr, _fused_adamw.tensor_lr_out
|
15595
15672
|
|
15596
15673
|
- func: _fused_sgd_(Tensor(a!)[] self, Tensor(b!)[] grads, Tensor(c!)[] momentum_buffer_list, *, float weight_decay, float momentum, float lr, float dampening, bool nesterov, bool maximize, bool is_first_step, Tensor? grad_scale=None, Tensor? found_inf=None) -> ()
|
@@ -15599,6 +15676,7 @@
|
|
15599
15676
|
dispatch:
|
15600
15677
|
CPU: _fused_sgd_kernel_cpu_
|
15601
15678
|
CUDA: _fused_sgd_kernel_cuda_
|
15679
|
+
MPS: _fused_sgd_kernel_mps_
|
15602
15680
|
autogen: _fused_sgd, _fused_sgd.out
|
15603
15681
|
|
15604
15682
|
- func: _fused_sgd_.tensor_lr(Tensor(a!)[] self, Tensor(b!)[] grads, Tensor(c!)[] momentum_buffer_list, *, float weight_decay, float momentum, Tensor lr, float dampening, bool nesterov, bool maximize, bool is_first_step, Tensor? grad_scale=None, Tensor? found_inf=None) -> ()
|
@@ -15609,6 +15687,7 @@
|
|
15609
15687
|
dispatch:
|
15610
15688
|
CPU: _fused_sgd_kernel_cpu_
|
15611
15689
|
CUDA: _fused_sgd_kernel_cuda_
|
15690
|
+
MPS: _fused_sgd_kernel_mps_
|
15612
15691
|
autogen: _fused_sgd.tensor_lr, _fused_sgd.tensor_lr_out
|
15613
15692
|
|
15614
15693
|
- func: _fused_adagrad_(Tensor(a!)[] self, Tensor(b!)[] grads, Tensor(c!)[] state_sums, Tensor(d!)[] state_steps, *, float lr, float lr_decay, float weight_decay, float eps, bool maximize, Tensor? grad_scale=None, Tensor? found_inf=None) -> ()
|