..
cuda_attention_kvcache.h
[feature] add fused attention_kvcache operator support ( #179 )
2023-11-14 23:44:22 +08:00
cuda_clip.h
Dev for 202303ddl ( #66 )
2023-04-18 15:10:33 +08:00
cuda_common.h
tensor parallel for transformer ( #125 )
2023-09-14 14:19:45 +08:00
cuda_element_wise.h
support Dynamic tensor infer shape and fix memory pool ( #176 )
2023-11-23 13:11:50 +08:00
cuda_expand.h
"modified where" ( #131 )
2023-09-14 10:45:57 +08:00
cuda_kernel_wihtout_config.h
ADD: batch norm operator and cuda kernel. ( #44 )
2022-10-15 16:29:28 +08:00
cuda_layernorm.h
Add layer normalization ( #181 )
2023-11-24 15:15:14 +08:00
cuda_pad_slice.h
support Dynamic tensor infer shape and fix memory pool ( #176 )
2023-11-23 13:11:50 +08:00
cuda_runtime.h
impl distributed launch with NCCL ( #106 )
2023-09-05 09:47:35 +08:00
cuda_softmax.h
Cuda softmax ( #129 )
2023-11-06 08:56:23 +08:00
cuda_split_concat.h
support 8D tensor, add test example ( #170 )
2023-10-31 10:47:36 +08:00
cuda_transpose.h
Add cuda transpose kernel ( #115 )
2023-08-22 14:22:15 +08:00
cuda_unary.h
Add HardSigmoid and HardSwish ( #156 )
2023-10-10 22:41:06 +08:00
cuda_utility.h
Simplify tensor transfer between CPU and CUDA ( #10 )
2022-08-25 11:29:16 +08:00
cuda_where.h
add infer index function ( #175 )
2023-11-24 09:24:25 +08:00
gather.h
Add GatherElements op and cuda kernel ( #149 )
2023-10-12 09:18:12 +08:00
gbmm_g2bmm.cuh
Fix CMake USE_CUDA ( #36 )
2022-09-21 12:28:00 +08:00
gbmm_g2bmm.h
Fix CMake USE_CUDA ( #36 )
2022-09-21 12:28:00 +08:00
nccl_communicator.h
impl distributed launch with NCCL ( #106 )
2023-09-05 09:47:35 +08:00
operator_timer.h
ADD: batch norm operator and cuda kernel. ( #44 )
2022-10-15 16:29:28 +08:00
resize.cuh
ADD: reconfig ResizeObj, support "tf_crop_and_resize " and cubic coeff kernel. ( #59 )
2022-12-24 04:02:21 +08:00