InfiniTensor/test/operators
constroy Li f60767a770
impl distributed launch with NCCL (#106)
* add cmake bits about NCCL

* move example to examples/NNmodel

* impl NCCL communicator

* add comm related function to Runtime

* export runtime interface

* add launch.py

* use unique name to distingush the the NCCL ID file

* add timeout to communicator init

* expose communicator obj from runtime obj, add unit test for nccl communicator

* reformat files

* Add allReduce operator and cuda nccl allReduce kernel

* impl model parallel for resnet

* add allGather nccl kernel and operator

* Add allreduce allgather operator tests, change allgather kernel to output list of tensor, fix shape infer, handle nullptr output

* fix format of onnx.py

* use concat following AllGather

* get tensor parallel for resnet

* fix format of graph_handler.cc

* change BUILD_DIST default to OFF

* polish code of communicator

* update .gitignore

* Add broadcast operator and cuda kernel

* Add comments for operators

* remove const of class member

* move communicator to CudaRuntimeObj

* Add an empty line at EOF.

---------

Co-authored-by: panzezhong <panzezhong@qiyuanlab.com>
Co-authored-by: Haojie Wang <haojie0429@gmail.com>
2023-09-05 09:47:35 +08:00
..
test_all_gather.cc impl distributed launch with NCCL (#106) 2023-09-05 09:47:35 +08:00
test_all_reduce.cc impl distributed launch with NCCL (#106) 2023-09-05 09:47:35 +08:00
test_batch_norm.cc refactor(core): 添加新的 `OpType` 定义 (#99) 2023-08-07 11:17:05 +08:00
test_broadcast.cc impl distributed launch with NCCL (#106) 2023-09-05 09:47:35 +08:00
test_clip.cc memory_allocator (#103) 2023-08-13 13:39:35 +08:00
test_concat.cc ADD: add mkl runtime for intel cpu , and add mkl kernel for matmul/conv/convtransposed. (#61) 2023-03-27 21:28:49 +08:00
test_conv.cc ADD: add mkl runtime for intel cpu , and add mkl kernel for matmul/conv/convtransposed. (#61) 2023-03-27 21:28:49 +08:00
test_conv_transposed_2d.cc ADD: add mkl runtime for intel cpu , and add mkl kernel for matmul/conv/convtransposed. (#61) 2023-03-27 21:28:49 +08:00
test_element_wise.cc ADD: add mkl runtime for intel cpu , and add mkl kernel for matmul/conv/convtransposed. (#61) 2023-03-27 21:28:49 +08:00
test_expand.cc 框架支持bert/gpt2模型构图 (#94) 2023-08-29 16:06:52 +08:00
test_extend.cc ADD: add mkl runtime for intel cpu , and add mkl kernel for matmul/conv/convtransposed. (#61) 2023-03-27 21:28:49 +08:00
test_gather.cc 框架支持bert/gpt2模型构图 (#94) 2023-08-29 16:06:52 +08:00
test_matmul.cc support mixed dtype (#102) 2023-08-16 21:49:43 +08:00
test_pad.cc ADD: add mkl runtime for intel cpu , and add mkl kernel for matmul/conv/convtransposed. (#61) 2023-03-27 21:28:49 +08:00
test_pooling.cc ADD: add mkl runtime for intel cpu , and add mkl kernel for matmul/conv/convtransposed. (#61) 2023-03-27 21:28:49 +08:00
test_reduce_mean.cc 框架支持bert/gpt2模型构图 (#94) 2023-08-29 16:06:52 +08:00
test_reshape.cc Cpu backend2 (#77) 2023-04-17 12:15:23 +08:00
test_resize.cc ADD: add mkl runtime for intel cpu , and add mkl kernel for matmul/conv/convtransposed. (#61) 2023-03-27 21:28:49 +08:00
test_slice.cc Dev for 202303ddl (#66) 2023-04-18 15:10:33 +08:00
test_split.cc impl sqrt on CUDA (#109) 2023-08-18 12:17:47 +08:00
test_transpose.cc support mixed dtype (#102) 2023-08-16 21:49:43 +08:00
test_where.cc 框架支持bert/gpt2模型构图 (#94) 2023-08-29 16:06:52 +08:00