forked from jiuyuan/InfiniTensor
51086d2b8d
* - Remove dataType from the kernel registration. * - support fp16 for conv * - cpu kernel: adapt the new registration mechanism * modified all register kernel * add where fp16 * add layernorm fp16 * add split_concat fp16 * - element_wise support fp16 * feat: support transpose fp16 * feat: support sliceOp fp16 * - unary support fp16 * - feat: support reduceOp fp16 * feat: support matmulOp/expandOp fp16 * feat: support powOp int8 * add cuda cast & support half-precision for gather * style: fix style * feat:support int8 for gather * style:fix style * modified test_cuda_conv_transposed * fix: fix dist code to support fp16 * fix(graph.cc): fix topo_sort * fix: fix recv and send kernel registration * feat: add field tensors for stub * refactor(frontend): 先排序后构图 Signed-off-by: YdrMaster <ydrml@hotmail.com> * fix: 为中间结果提供tensor到node的mapping * fix (slice): add guard for area out of range * fix: fix matmul fp16 * fix: fix re-dataMalloc for weight tensor and use of naive allocator * feat: add dataType filter for cuda kernel * feat: bang kernel adapt the new registration mechanism * fix: fix some error on mlu * feat: intelcpu kernel adapt the new registration mechanism * feat: modify kernel registration on kunlun * fix intelcpu compiler bug * feat: bang reshape support all dataType * fix: fix bang reduce * fix(all_reduce.cc): fix as reviewer suggessted * fix: fix style and restore unary test codes --------- Signed-off-by: YdrMaster <ydrml@hotmail.com> Co-authored-by: xgqdut2016 <kenan_gewei@163.com> Co-authored-by: xgqdut2016 <140036308+xgqdut2016@users.noreply.github.com> Co-authored-by: zhangyunze <z13785159769@163.com> Co-authored-by: OdinaryWord <sx-hz@163.com> Co-authored-by: YdrMaster <ydrml@hotmail.com> Co-authored-by: panzezhong <panzezhong@qiyuanlab.com> |
||
---|---|---|
.github/workflows | ||
3rd-party | ||
cmake | ||
docs | ||
examples | ||
include | ||
proto | ||
pyinfinitensor | ||
python | ||
scripts | ||
src | ||
test | ||
.clang-format | ||
.cmake-format.json | ||
.gitignore | ||
.gitmodules | ||
CHANGELOG.md | ||
CMakeLists.txt | ||
LICENSE | ||
Makefile | ||
README.md | ||
README_CN.md | ||
env.sh |
README.md
InfiniTensor
InfiniTensor is a high-performance inference engine tailored for GPUs and AI accelerators. Its design focuses on effective deployment and swift academic validation.
Get started
Make Commands
make
/make build
: Builds the project;make install-python
: Builds the project then install the python frontend;make test-cpp
: Builds the project then run cpp unit tests;make test-onnx
: Run python unit tests;
- Sets env:
TEST=OFF
to accelerate compiling.- Sets env:
CUDA=ON
to enable cuda.- Sets env:
BANG=ON
to enable bang.
CMake Options
There are several configurable CMake options, see the CMakeLists.txt file.
- If
USE_BACKTRACE
isON
,libdw-dev
have to be installed. See the README of backward-cpp for details. - If
USE_PROTOBUF
isON
,protobuf
have to be installed. See the README of protobuf for details. - If
USE_CUDA
isON
,cuda
have to be installed.
Roadmap
- EinNet is going to be merged into the main branch.
- Integration of PET, a tensor program optimizer supporting partially equivalent transformations.
- Supported hardware
- ✔ NVIDIA GPU
- ✔ Cambricon MLU
- ⬜ Ascend NPU
- ⬜ Kunlunxin XPU
Contributor Guide
InfiniTensor development is based on the pull request on Github. Before requesting for merging, a PR should satisfy the following requirements
- Pass all tests.
- Now CI on Github will test everything that can be tested in the ci environment, including code format. So, script
test/script/clang_format_inplace.sh
is for formatting all code. - Contributors should run
ctest
manually and copy its output to the PR. Use fenced code blocks (triple backquotes, i.e.,```
) to avoid referencing in Github. Otherwise,#
in the output is interpreted as a Github reference. Do not directly paste the ctest output in commit messages either for the same reason.
- Now CI on Github will test everything that can be tested in the ci environment, including code format. So, script
- Receive at least one approval from reviewers.
- PR title should be concise since it is going to be the commit message in the main branch after merging and squashing.
Reference
Please cite EinNet or PET in your publications if it helps your research:
@article{zheng2023einnet,
title={EINNET: Optimizing Tensor Programs with Derivation-Based Transformations},
author={Zheng, Liyan and Wang, Haojie and Zhai, Jidong and Hu, Muyan and Ma, Zixuan and Wang, Tuowei and Huang, Shuhong and Miao, Xupeng and Tang, Shizhi and Huang, Kezhao and Jia, Zhihao},
booktitle={17th USENIX Symposium on Operating Systems Design and Implementation (OSDI 23)},
pages={739--755},
year={2023}
}
@inproceedings{wang2021pet,
title={PET: Optimizing tensor programs with partially equivalent transformations and automated corrections},
author={Wang, Haojie and Zhai, Jidong and Gao, Mingyu and Ma, Zixuan and Tang, Shizhi and Zheng, Liyan and Li, Yuanzhi and Rong, Kaiyuan and Chen, Yuanyong and Jia, Zhihao},
booktitle={15th USENIX Symposium on Operating Systems Design and Implementation (OSDI 21)},
pages={37--54},
year={2021}
}