forked from jiuyuan/InfiniTensor
48ec730579
* add cmake bits about NCCL * move example to examples/NNmodel * impl NCCL communicator * add comm related function to Runtime * export runtime interface * add launch.py * use unique name to distingush the the NCCL ID file * add timeout to communicator init * expose communicator obj from runtime obj, add unit test for nccl communicator * reformat files * Add allReduce operator and cuda nccl allReduce kernel * impl model parallel for resnet * add allGather nccl kernel and operator * Add allreduce allgather operator tests, change allgather kernel to output list of tensor, fix shape infer, handle nullptr output * fix format of onnx.py * use concat following AllGather * get tensor parallel for resnet * fix format of graph_handler.cc * change BUILD_DIST default to OFF * polish code of communicator * update .gitignore * export min/max to python * fix MatMul * modify launch.py to run opt * hack to treat ReduceSum as AllReduceSum * throw exception in cuda error * fix parallel_opt.py * improve the error prompt and cuda error check * fix GatherObj::GatherObj member init * fix size calculation for scalar (rank = 0) tensor * MatMul supports bias * fix add bias for row parallel gemm * add --gen_std to launch.py * fix AllReduceNCCL * update launch.py * less log * update parallel_opt * update launch.py * add __eq__ for Placement sub-classes * less benchmark run * fix placement infer for matmul * fix vacabuary size * fix Exception * Add shard tensor with group to support gpt2 * Add find successor function to find split op at different depth * recover CommunicatorObj * improve error mesasge * optimize parallel_opt.py * optimize launch.py * recover docs for all_reduce and all_gather * - support concat for kvcache * - modify allocator * - add tensorType - modify allocator to support memory allocation based on tensorType * - fix allocator init * - support kvcache by running 2 stub distributively * - fix name * - remove unused flag * - fix wrong pb name * - fix as constroy suggessed * - fix launch.py format --------- Co-authored-by: constroy <constroy.li@gmail.com> Co-authored-by: panzezhong <panzezhong@qiyuanlab.com> |
||
---|---|---|
.github/workflows | ||
3rd-party | ||
cmake | ||
docs | ||
examples | ||
include | ||
proto | ||
pyinfinitensor | ||
python | ||
scripts | ||
src | ||
test | ||
.clang-format | ||
.cmake-format.json | ||
.gitignore | ||
.gitmodules | ||
CHANGELOG.md | ||
CMakeLists.txt | ||
LICENSE | ||
Makefile | ||
README.md | ||
README_CN.md | ||
env.sh |
README.md
InfiniTensor
InfiniTensor is a high-performance inference engine tailored for GPUs and AI accelerators. Its design focuses on effective deployment and swift academic validation.
Get started
Make Commands
make
/make build
: Builds the project;make install-python
: Builds the project then install the python frontend;make test-cpp
: Builds the project then run cpp unit tests;make test-onnx
: Run python unit tests;
- Sets env:
TEST=OFF
to accelerate compiling.- Sets env:
CUDA=ON
to enable cuda.- Sets env:
BANG=ON
to enable bang.
CMake Options
There are several configurable CMake options, see the CMakeLists.txt file.
- If
USE_BACKTRACE
isON
,libdw-dev
have to be installed. See the README of backward-cpp for details. - If
USE_PROTOBUF
isON
,protobuf
have to be installed. See the README of protobuf for details. - If
USE_CUDA
isON
,cuda
have to be installed.
Roadmap
- EinNet is going to be merged into the main branch.
- Integration of PET, a tensor program optimizer supporting partially equivalent transformations.
- Supported hardware
- ✔ NVIDIA GPU
- ✔ Cambricon MLU
- ⬜ Ascend NPU
- ⬜ Kunlunxin XPU
Contributor Guide
InfiniTensor development is based on the pull request on Github. Before requesting for merging, a PR should satisfy the following requirements
- Pass all tests.
- Now CI on Github will test everything that can be tested in the ci environment, including code format. So, script
test/script/clang_format_inplace.sh
is for formatting all code. - Contributors should run
ctest
manually and copy its output to the PR. Use fenced code blocks (triple backquotes, i.e.,```
) to avoid referencing in Github. Otherwise,#
in the output is interpreted as a Github reference. Do not directly paste the ctest output in commit messages either for the same reason.
- Now CI on Github will test everything that can be tested in the ci environment, including code format. So, script
- Receive at least one approval from reviewers.
- PR title should be concise since it is going to be the commit message in the main branch after merging and squashing.
Reference
Please cite EinNet or PET in your publications if it helps your research:
@article{zheng2023einnet,
title={EINNET: Optimizing Tensor Programs with Derivation-Based Transformations},
author={Zheng, Liyan and Wang, Haojie and Zhai, Jidong and Hu, Muyan and Ma, Zixuan and Wang, Tuowei and Huang, Shuhong and Miao, Xupeng and Tang, Shizhi and Huang, Kezhao and Jia, Zhihao},
booktitle={17th USENIX Symposium on Operating Systems Design and Implementation (OSDI 23)},
pages={739--755},
year={2023}
}
@inproceedings{wang2021pet,
title={PET: Optimizing tensor programs with partially equivalent transformations and automated corrections},
author={Wang, Haojie and Zhai, Jidong and Gao, Mingyu and Ma, Zixuan and Tang, Shizhi and Zheng, Liyan and Li, Yuanzhi and Rong, Kaiyuan and Chen, Yuanyong and Jia, Zhihao},
booktitle={15th USENIX Symposium on Operating Systems Design and Implementation (OSDI 21)},
pages={37--54},
year={2021}
}