InfiniTensor/include/operators
Hardy 8b61b0e397
support ascend (#165)
* fix

* fix code

* fix format

* fix format

* fix

* fix

* addAbs

* more Unary

* add kernels

* fix concat&pooling test code

* add softmax/element_wise kernel

* fix format

* add reshape

* support for llama

* add maxpooling & flatten

* add conv_transpose&&native maxpooling

* add conv_transpose

* add communication operator

* fix

* style: fix format

* style: fix format

* add depthTospace&&resize

* add layernorm

* format

* add gemm

* add leakyRelu op

* modified format

* modified onnx leakyrelu alpha

* modified batchnorm

* fix gemm & avgpooling

* fix: onnx resize op input is none bug

* add instancenorm, use layernorm replace instance, error

* modiefied format, replace layernorm as instancenorm

* fix: onnx resize op input is none bug

* add pad2d kernel

* modified format

* fix op

* fix resize

* remove sync in op

* Update INSTALL_GUIDE_CN.md for ASCEND

* Update env.sh

* format

* fix test_resize

* fix resize

* fix test_resize_

* fix test_resize_

* add HcclCommDestroy && use default context

* install onnxtuntime

* install onnx-simplifier

* install numpy

* fix bug after merge

* remove CHECK_RET&LOG_PRINT

* fix test_ascend_layernorm

* fix test_cuda_resize

* fix test_ascend_*

* fix format

---------

Co-authored-by: Haojie Wang <haojie0429@gmail.com>
Co-authored-by: wanghailu <wanghailu@qiyuanlab.com>
Co-authored-by: OdinaryWord <sx-hz@163.com>
Co-authored-by: xgqdut2016 <kenan_gewei@163.com>
Co-authored-by: zhangyunze <z13785159769@163.com>
Co-authored-by: Songxin <sx-hz@hotmail.com>
Co-authored-by: zhangyue <138768300+zhangyue207@users.noreply.github.com>
Co-authored-by: zhangyue <zhangyue@qiyuanlab.com>
Co-authored-by: sx941227 <14507528+sx941227@user.noreply.gitee.com>
Co-authored-by: zhangyunze <zhangyunze@qiyuanlab.com>
Co-authored-by: Chenjie Duan <44265800+kilinchange@users.noreply.github.com>
2024-08-20 22:09:33 +08:00
..
G2BMM.h support Dynamic tensor infer shape and fix memory pool (#176) 2023-11-23 13:11:50 +08:00
GBMM.h support Dynamic tensor infer shape and fix memory pool (#176) 2023-11-23 13:11:50 +08:00
activation_backward.h support Dynamic tensor infer shape and fix memory pool (#176) 2023-11-23 13:11:50 +08:00
all_gather.h support Dynamic tensor infer shape and fix memory pool (#176) 2023-11-23 13:11:50 +08:00
all_reduce.h support Dynamic tensor infer shape and fix memory pool (#176) 2023-11-23 13:11:50 +08:00
attention_kvcache.h use workspace to optimize kvcache attention 2024-01-25 10:33:01 +08:00
batch_norm.h support Dynamic tensor infer shape and fix memory pool (#176) 2023-11-23 13:11:50 +08:00
broadcast.h support Dynamic tensor infer shape and fix memory pool (#176) 2023-11-23 13:11:50 +08:00
concat.h support Dynamic tensor infer shape and fix memory pool (#176) 2023-11-23 13:11:50 +08:00
conv.h Add `Conv3d` operator and its naive CPU kernel implementation 2024-06-25 14:41:21 +08:00
det.h support Dynamic tensor infer shape and fix memory pool (#176) 2023-11-23 13:11:50 +08:00
dropout.h support Dynamic tensor infer shape and fix memory pool (#176) 2023-11-23 13:11:50 +08:00
element_wise.h support Dynamic tensor infer shape and fix memory pool (#176) 2023-11-23 13:11:50 +08:00
expand.h support Dynamic tensor infer shape and fix memory pool (#176) 2023-11-23 13:11:50 +08:00
extend.h support Dynamic tensor infer shape and fix memory pool (#176) 2023-11-23 13:11:50 +08:00
gather.h support Dynamic tensor infer shape and fix memory pool (#176) 2023-11-23 13:11:50 +08:00
instance_norm.h support ascend (#165) 2024-08-20 22:09:33 +08:00
layer_norm.h Add layer normalization (#181) 2023-11-24 15:15:14 +08:00
lrn.h Fix bang (#198) 2023-12-28 13:44:10 +08:00
matmul.h feature: add parameter to config matmul compute type (#218) 2024-03-26 09:00:45 +08:00
membound.h support Dynamic tensor infer shape and fix memory pool (#176) 2023-11-23 13:11:50 +08:00
pad.h support Dynamic tensor infer shape and fix memory pool (#176) 2023-11-23 13:11:50 +08:00
pooling.h support Dynamic tensor infer shape and fix memory pool (#176) 2023-11-23 13:11:50 +08:00
recv.h Add send and recv operators based on NCCL (#182) 2023-12-14 16:38:03 +08:00
reduce.h Add ReduceSum op and kernel (#160) 2023-11-24 09:29:58 +08:00
reshape.h support Dynamic tensor infer shape and fix memory pool (#176) 2023-11-23 13:11:50 +08:00
resize.h support ascend (#165) 2024-08-20 22:09:33 +08:00
rms_norm.h Accelerate llama (#219) 2024-04-01 08:46:05 +08:00
rope.h add comment for rope operator 2024-02-04 10:57:01 +08:00
send.h Add send and recv operators based on NCCL (#182) 2023-12-14 16:38:03 +08:00
slice.h support Dynamic tensor infer shape and fix memory pool (#176) 2023-11-23 13:11:50 +08:00
softmax.h support Dynamic tensor infer shape and fix memory pool (#176) 2023-11-23 13:11:50 +08:00
split.h support Dynamic tensor infer shape and fix memory pool (#176) 2023-11-23 13:11:50 +08:00
squeeze.h 解除前端对onnx infershape功能的依赖 (#206) 2024-01-12 14:54:27 +08:00
transpose.h support Dynamic tensor infer shape and fix memory pool (#176) 2023-11-23 13:11:50 +08:00
unary.h Add ELU operator (#237) 2024-07-07 17:35:03 +08:00
unsqueeze.h 解除前端对onnx infershape功能的依赖 (#206) 2024-01-12 14:54:27 +08:00
where.h support Dynamic tensor infer shape and fix memory pool (#176) 2023-11-23 13:11:50 +08:00