@@ -6187,247 +6187,160 @@
6187
6187
CUDA : foreach_tensor_add_scalar_kernel_cuda
6188
6188
6189
6189
- func : _foreach_add_.Scalar(Tensor(a!)[] self, Scalar scalar) -> ()
6190
- use_c10_dispatcher : full
6191
6190
device_guard : False
6192
6191
variants : function
6193
6192
dispatch :
6194
6193
CPU : foreach_tensor_add_scalar_kernel_slow_
6195
6194
CUDA : foreach_tensor_add_scalar_kernel_cuda_
6196
6195
6197
6196
- func : _foreach_sub.Scalar(Tensor[] tensors, Scalar scalar) -> Tensor[]
6198
- use_c10_dispatcher : full
6199
6197
device_guard : False
6200
6198
variants : function
6201
6199
dispatch :
6202
6200
CPU : foreach_tensor_sub_scalar_kernel_slow
6203
6201
CUDA : foreach_tensor_sub_scalar_kernel_cuda
6204
6202
6205
6203
- func : _foreach_sub_.Scalar(Tensor(a!)[] self, Scalar scalar) -> ()
6206
- use_c10_dispatcher : full
6207
6204
device_guard : False
6208
6205
variants : function
6209
6206
dispatch :
6210
6207
CPU : foreach_tensor_sub_scalar_kernel_slow_
6211
6208
CUDA : foreach_tensor_sub_scalar_kernel_cuda_
6212
6209
6213
6210
- func : _foreach_mul.Scalar(Tensor[] tensors, Scalar scalar) -> Tensor[]
6214
- use_c10_dispatcher : full
6215
6211
device_guard : False
6216
6212
variants : function
6217
6213
dispatch :
6218
6214
CPU : foreach_tensor_mul_scalar_kernel_slow
6219
6215
CUDA : foreach_tensor_mul_scalar_kernel_cuda
6220
6216
6221
6217
- func : _foreach_mul_.Scalar(Tensor(a!)[] self, Scalar scalar) -> ()
6222
- use_c10_dispatcher : full
6223
6218
device_guard : False
6224
6219
variants : function
6225
6220
dispatch :
6226
6221
CPU : foreach_tensor_mul_scalar_kernel_slow_
6227
6222
CUDA : foreach_tensor_mul_scalar_kernel_cuda_
6228
6223
6229
6224
- func : _foreach_div.Scalar(Tensor[] tensors, Scalar scalar) -> Tensor[]
6230
- use_c10_dispatcher : full
6231
6225
device_guard : False
6232
6226
variants : function
6233
6227
dispatch :
6234
6228
CPU : foreach_tensor_div_scalar_kernel_slow
6235
6229
CUDA : foreach_tensor_div_scalar_kernel_cuda
6236
6230
6237
6231
- func : _foreach_div_.Scalar(Tensor(a!)[] self, Scalar scalar) -> ()
6238
- use_c10_dispatcher : full
6239
6232
device_guard : False
6240
6233
variants : function
6241
6234
dispatch :
6242
6235
CPU : foreach_tensor_div_scalar_kernel_slow_
6243
6236
CUDA : foreach_tensor_div_scalar_kernel_cuda_
6244
6237
6245
- - func : _foreach_add.List(Tensor[] tensors1, Tensor[] tensors2, *, Scalar alpha=1) -> Tensor[]
6246
- use_c10_dispatcher : full
6238
+ - func : _foreach_add.List(Tensor[] tensors1, Tensor[] tensors2, Scalar alpha=1) -> Tensor[]
6247
6239
device_guard : False
6248
6240
variants : function
6249
6241
dispatch :
6250
6242
CPU : foreach_tensor_add_list_kernel_slow
6251
6243
CUDA : foreach_tensor_add_list_kernel_cuda
6252
6244
6253
- - func : _foreach_add_.List(Tensor(a!)[] self, Tensor[] other, *, Scalar alpha=1) -> ()
6254
- use_c10_dispatcher : full
6245
+ - func : _foreach_add_.List(Tensor(a!)[] self, Tensor[] other, Scalar alpha=1) -> ()
6255
6246
device_guard : False
6256
6247
variants : function
6257
6248
dispatch :
6258
6249
CPU : foreach_tensor_add_list_kernel_slow_
6259
6250
CUDA : foreach_tensor_add_list_kernel_cuda_
6260
6251
6261
- - func : _foreach_sub.List(Tensor[] tensors1, Tensor[] tensors2, *, Scalar alpha=1) -> Tensor[]
6262
- use_c10_dispatcher : full
6252
+ - func : _foreach_sub.List(Tensor[] tensors1, Tensor[] tensors2, Scalar alpha=1) -> Tensor[]
6263
6253
device_guard : False
6264
6254
variants : function
6265
6255
dispatch :
6266
6256
CPU : foreach_tensor_sub_list_kernel_slow
6267
6257
CUDA : foreach_tensor_sub_list_kernel_cuda
6268
6258
6269
- - func : _foreach_sub_.List(Tensor(a!)[] self, Tensor[] other, *, Scalar alpha=1) -> ()
6270
- use_c10_dispatcher : full
6259
+ - func : _foreach_sub_.List(Tensor(a!)[] self, Tensor[] other, Scalar alpha=1) -> ()
6271
6260
device_guard : False
6272
6261
variants : function
6273
6262
dispatch :
6274
6263
CPU : foreach_tensor_sub_list_kernel_slow_
6275
6264
CUDA : foreach_tensor_sub_list_kernel_cuda_
6276
6265
6277
6266
- func : _foreach_mul.List(Tensor[] tensors1, Tensor[] tensors2) -> Tensor[]
6278
- use_c10_dispatcher : full
6279
6267
device_guard : False
6280
6268
variants : function
6281
6269
dispatch :
6282
6270
CPU : foreach_tensor_mul_list_kernel_slow
6283
6271
CUDA : foreach_tensor_mul_list_kernel_cuda
6284
6272
6285
6273
- func : _foreach_mul_.List(Tensor(a!)[] self, Tensor[] other) -> ()
6286
- use_c10_dispatcher : full
6287
6274
device_guard : False
6288
6275
variants : function
6289
6276
dispatch :
6290
6277
CPU : foreach_tensor_mul_list_kernel_slow_
6291
6278
CUDA : foreach_tensor_mul_list_kernel_cuda_
6292
6279
6293
- - func : _foreach_div.List(Tensor[] tensors1, Tensor[] tensors2) -> Tensor[]
6294
- use_c10_dispatcher : full
6280
+ - func : _foreach_div.List(Tensor(a!)[] self, Tensor[] other) -> Tensor[]
6295
6281
device_guard : False
6296
6282
variants : function
6297
6283
dispatch :
6298
6284
CPU : foreach_tensor_div_list_kernel_slow
6299
6285
CUDA : foreach_tensor_div_list_kernel_cuda
6300
6286
6301
6287
- func : _foreach_div_.List(Tensor(a!)[] self, Tensor[] other) -> ()
6302
- use_c10_dispatcher : full
6303
6288
device_guard : False
6304
6289
variants : function
6305
6290
dispatch :
6306
6291
CPU : foreach_tensor_div_list_kernel_slow_
6307
6292
CUDA : foreach_tensor_div_list_kernel_cuda_
6308
6293
6309
- - func : _foreach_add.ScalarList(Tensor[] tensors, float[] scalars) -> Tensor[]
6310
- use_c10_dispatcher : full
6311
- device_guard : False
6312
- variants : function
6313
- dispatch :
6314
- CPU : foreach_tensor_add_scalarlist_kernel_slow
6315
- CUDA : foreach_tensor_add_scalarlist_kernel_cuda
6316
-
6317
- - func : _foreach_add_.ScalarList(Tensor(a!)[] self, float[] scalars) -> ()
6318
- use_c10_dispatcher : full
6319
- device_guard : False
6320
- variants : function
6321
- dispatch :
6322
- CPU : foreach_tensor_add_scalarlist_kernel_slow_
6323
- CUDA : foreach_tensor_add_scalarlist_kernel_cuda_
6324
-
6325
- - func : _foreach_sub.ScalarList(Tensor[] tensors, float[] scalars) -> Tensor[]
6326
- use_c10_dispatcher : full
6327
- device_guard : False
6328
- variants : function
6329
- dispatch :
6330
- CPU : foreach_tensor_sub_scalarlist_kernel_slow
6331
- CUDA : foreach_tensor_sub_scalarlist_kernel_cuda
6332
-
6333
- - func : _foreach_sub_.ScalarList(Tensor(a!)[] self, float[] scalars) -> ()
6334
- use_c10_dispatcher : full
6335
- device_guard : False
6336
- variants : function
6337
- dispatch :
6338
- CPU : foreach_tensor_sub_scalarlist_kernel_slow_
6339
- CUDA : foreach_tensor_sub_scalarlist_kernel_cuda_
6340
-
6341
- - func : _foreach_div.ScalarList(Tensor[] tensors, float[] scalars) -> Tensor[]
6342
- use_c10_dispatcher : full
6343
- device_guard : False
6344
- variants : function
6345
- dispatch :
6346
- CPU : foreach_tensor_div_scalarlist_kernel_slow
6347
- CUDA : foreach_tensor_div_scalarlist_kernel_cuda
6348
-
6349
- - func : _foreach_div_.ScalarList(Tensor(a!)[] self, float[] scalars) -> ()
6350
- use_c10_dispatcher : full
6351
- device_guard : False
6352
- variants : function
6353
- dispatch :
6354
- CPU : foreach_tensor_div_scalarlist_kernel_slow_
6355
- CUDA : foreach_tensor_div_scalarlist_kernel_cuda_
6356
-
6357
- - func : _foreach_mul.ScalarList(Tensor[] tensors, float[] scalars) -> Tensor[]
6358
- use_c10_dispatcher : full
6359
- device_guard : False
6360
- variants : function
6361
- dispatch :
6362
- CPU : foreach_tensor_mul_scalarlist_kernel_slow
6363
- CUDA : foreach_tensor_mul_scalarlist_kernel_cuda
6364
-
6365
- - func : _foreach_mul_.ScalarList(Tensor(a!)[] self, float[] scalars) -> ()
6366
- use_c10_dispatcher : full
6367
- device_guard : False
6368
- variants : function
6369
- dispatch :
6370
- CPU : foreach_tensor_mul_scalarlist_kernel_slow_
6371
- CUDA : foreach_tensor_mul_scalarlist_kernel_cuda_
6372
-
6373
6294
- func : _foreach_exp(Tensor[] tensors) -> Tensor[]
6374
- use_c10_dispatcher : full
6375
6295
device_guard : False
6376
6296
variants : function
6377
6297
dispatch :
6378
6298
CPU : foreach_tensor_exp_slow
6379
6299
CUDA : foreach_tensor_exp_cuda
6380
6300
6381
6301
- func : _foreach_exp_(Tensor(a!)[] self) -> ()
6382
- use_c10_dispatcher : full
6383
6302
device_guard : False
6384
6303
variants : function
6385
6304
dispatch :
6386
6305
CPU : foreach_tensor_exp_slow_
6387
6306
CUDA : foreach_tensor_exp_cuda_
6388
6307
6389
6308
- func : _foreach_sqrt(Tensor[] tensors) -> Tensor[]
6390
- use_c10_dispatcher : full
6391
6309
device_guard : False
6392
6310
variants : function
6393
6311
dispatch :
6394
6312
CPU : foreach_tensor_sqrt_slow
6395
6313
CUDA : foreach_tensor_sqrt_cuda
6396
6314
6397
6315
- func : _foreach_sqrt_(Tensor(a!)[] self) -> ()
6398
- use_c10_dispatcher : full
6399
6316
device_guard : False
6400
6317
variants : function
6401
6318
dispatch :
6402
6319
CPU : foreach_tensor_sqrt_slow_
6403
6320
CUDA : foreach_tensor_sqrt_cuda_
6404
6321
6405
6322
- func : _foreach_addcdiv_(Tensor(a!)[] self, Tensor[] tensor1, Tensor[] tensor2, Scalar value=1) -> ()
6406
- use_c10_dispatcher : full
6407
6323
device_guard : False
6408
6324
variants : function
6409
6325
dispatch :
6410
6326
CPU : foreach_tensor_addcdiv_slow_
6411
6327
CUDA : foreach_tensor_addcdiv_cuda_
6412
6328
6413
6329
- func : _foreach_addcmul_(Tensor(a!)[] self, Tensor[] tensor1, Tensor[] tensor2, Scalar value=1) -> ()
6414
- use_c10_dispatcher : full
6415
6330
device_guard : False
6416
6331
variants : function
6417
6332
dispatch :
6418
6333
CPU : foreach_tensor_addcmul_slow_
6419
6334
CUDA : foreach_tensor_addcmul_cuda_
6420
6335
6421
6336
- func : _foreach_addcdiv(Tensor[] input, Tensor[] tensor1, Tensor[] tensor2, Scalar value=1) -> Tensor[]
6422
- use_c10_dispatcher : full
6423
6337
device_guard : False
6424
6338
variants : function
6425
6339
dispatch :
6426
6340
CPU : foreach_tensor_addcdiv_slow
6427
6341
CUDA : foreach_tensor_addcdiv_cuda
6428
6342
6429
6343
- func : _foreach_addcmul(Tensor[] input, Tensor[] tensor1, Tensor[] tensor2, Scalar value=1) -> Tensor[]
6430
- use_c10_dispatcher : full
6431
6344
device_guard : False
6432
6345
variants : function
6433
6346
dispatch :
0 commit comments