Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merge master branch to resolve conflicts #571

Merged
merged 52 commits into from
Jun 8, 2022
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
52 commits
Select commit Hold shift + click to select a range
7cfe6e6
fix pose demo and windows build (#307)
irexyc Apr 2, 2022
442e9cd
add postprocessing_masks gpu version (#276)
xizi Apr 6, 2022
85c46ee
fixed a bug causes text-recognizer to fail when (non-NULL) empty bbox…
lzhangzz Apr 7, 2022
6e7e219
[Fix] include missing <type_traits> for formatter.h (#313)
lzhangzz Apr 11, 2022
d7adf81
[Fix] MMEditing cannot save results when testing (#336)
SingleZombie Apr 14, 2022
89ce8e2
docs(build): fix typo (#352)
tpoisonooo Apr 14, 2022
6aacede
fix openvino export tmp model, add binary flag (#353)
irexyc Apr 15, 2022
ade8e02
init circleci (#348)
lvhan028 Apr 15, 2022
fdbd3d1
fix wrong input mat type (#362)
irexyc Apr 15, 2022
88062e9
fix(docs): remove redundant doc tree (#360)
tpoisonooo Apr 15, 2022
b9c5487
fix missing ncnn_DIR & InferenceEngine_DIR (#364)
irexyc Apr 15, 2022
a8c75de
Fix mmdet openvino dynamic 300x300 cfg base (#372)
PeterH0323 Apr 18, 2022
957fd58
Fix: add onnxruntime building option in gpu dockerfile (#366)
Adenialzz Apr 18, 2022
9dad97e
Tutorial 03: torch2onnx (#365)
SingleZombie Apr 24, 2022
85f1778
[Docs] fix ncnn docs (#378)
hanrui1sensetime Apr 24, 2022
f9144f7
typo-fix (#397)
grimoire Apr 25, 2022
ee26593
add CUDA_TOOKIT_ROOT_DIR as tensorrt detect dir (#357)
irexyc Apr 25, 2022
f6fcee5
Fix docs (#398)
triple-Mu Apr 25, 2022
53ad86d
ort_net ONNX_TENSOR_ELEMENT_DATA_TYPE_BOOL (#383)
zly19540609 Apr 26, 2022
d9976c4
fix wrong buffer which will case onnxruntime-gpu crash with segmentai…
irexyc Apr 26, 2022
9560348
fix benchmark (#411)
RunningLeon Apr 27, 2022
8e6d4de
Add `sm_53` in cuda.cmake for Jetson Nano which will cashe when proce…
PeterH0323 Apr 27, 2022
cecd1ec
[Fix] fix feature test for `std::source_location` (#416)
lzhangzz Apr 27, 2022
72c19e9
fix format string (#417)
lzhangzz Apr 27, 2022
a9a4144
[Fix] Fix seg name (#394)
AllentDan Apr 27, 2022
21230d5
【Docs】Add ipython notebook tutorial (#234)
VVsssssk Apr 27, 2022
8aba06d
fix mmpose api (#396)
irexyc Apr 27, 2022
16ee9c7
[Fix] support latest spdlog (#423)
lzhangzz May 3, 2022
86ab063
fix pillarencode (#331)
VVsssssk May 5, 2022
5231e65
fix ONNXRuntime cuda test bug (#438)
NagatoYuki0943 May 7, 2022
c2f2edc
Fix ci in master branch (#441)
grimoire May 7, 2022
f45c1f0
[Doc] Improve Jetson tutorial install doc (#381)
PeterH0323 May 7, 2022
94148cb
Version comments added, torch install steps added. (#449)
tehkillerbee May 9, 2022
3786856
[Docs] Fix API documentation (#443)
zhouzaida May 9, 2022
0cd44a6
[Fix] Fix display bugs for windows (#451)
hanrui1sensetime May 12, 2022
21c2a85
[Docs] Minor fixes and translation of installation tutorial for Jetso…
Dchaoqun May 16, 2022
2f2ec27
If a cuda launch error occurs, verify if cuda device requires top_k t…
tehkillerbee May 17, 2022
ba641c3
[Fix] set optional arg a default value (#483)
AllentDan May 17, 2022
69111a6
Update: Optimize document (#484)
lakshanthad May 17, 2022
e057b87
fix: bbox_nms not onnxizing if batch size > 1 (#501)
gyf304 May 20, 2022
a4de9f3
change seperator of function marker (#499)
RunningLeon May 20, 2022
57baf21
[docs] Fix typo in tutorial (#509)
zhiqwang May 23, 2022
de3f18f
Fix docstring format (#495)
triple-Mu May 23, 2022
d16720b
Tutorial 04: onnx custom op (#508)
SingleZombie May 24, 2022
4f49763
fix mmseg twice resize (#480)
RunningLeon May 25, 2022
0878b8f
Fix mask test with mismatched device (#511)
RunningLeon May 26, 2022
32482e7
compat mmpose v0.26 (#518)
RunningLeon May 27, 2022
571b240
[Docs] adding new backends when using MMDeploy as a third package (#482)
AllentDan May 27, 2022
6fa1787
Tutorial 05: ONNX Model Editing (#517)
SingleZombie May 28, 2022
182cc51
fix pspnet torchscript conversion (#538)
RunningLeon Jun 1, 2022
2a0fcb6
changing the onnxwrapper script for gpu issue (#532)
sanjaypavo Jun 7, 2022
594c835
merge master and fix the conflicts
lvhan028 Jun 7, 2022
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion .pre-commit-config.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -39,7 +39,7 @@ repos:
rev: v2.1.0
hooks:
- id: codespell
args: ["--skip=third_party/*,*.proto"]
args: ["--skip=third_party/*,*.ipynb,*.proto"]

- repo: https://github.com/myint/docformatter
rev: v1.4
Expand Down
18 changes: 12 additions & 6 deletions csrc/backend_ops/tensorrt/common_impl/nms/allClassNMS.cu
Original file line number Diff line number Diff line change
Expand Up @@ -205,6 +205,18 @@ pluginStatus_t allClassNMS_gpu(cudaStream_t stream, const int num, const int num
(T_BBOX *)bbox_data, (T_SCORE *)beforeNMS_scores, (int *)beforeNMS_index_array,
(T_SCORE *)afterNMS_scores, (int *)afterNMS_index_array, flipXY);

cudaError_t code = cudaGetLastError();
if (code != cudaSuccess) {
// Verify if cuda dev0 requires top_k to be reduced;
// sm_53 (Jetson Nano) and sm_62 (Jetson TX2) requires reduced top_k < 1000
auto __cuda_arch__ = get_cuda_arch(0);
if ((__cuda_arch__ == 530 || __cuda_arch__ == 620) && top_k >= 1000) {
printf(
"Warning: pre_top_k need to be reduced for devices with arch 5.3, 6.2, got "
"pre_top_k=%d\n",
top_k);
}
}
CSC(cudaGetLastError(), STATUS_FAILURE);
return STATUS_SUCCESS;
}
Expand Down Expand Up @@ -243,13 +255,7 @@ pluginStatus_t allClassNMS(cudaStream_t stream, const int num, const int num_cla
const bool isNormalized, const DataType DT_SCORE, const DataType DT_BBOX,
void *bbox_data, void *beforeNMS_scores, void *beforeNMS_index_array,
void *afterNMS_scores, void *afterNMS_index_array, bool flipXY) {
auto __cuda_arch__ = get_cuda_arch(0); // assume there is only one arch 7.2 device
if (__cuda_arch__ == 720 && top_k >= 1000) {
printf("Warning: pre_top_k need to be reduced for devices with arch 7.2, got pre_top_k=%d\n",
top_k);
}
nmsLaunchConfigSSD lc(DT_SCORE, DT_BBOX);

for (unsigned i = 0; i < nmsFuncVec.size(); ++i) {
if (lc == nmsFuncVec[i]) {
DEBUG_PRINTF("all class nms kernel %d\n", i);
Expand Down
484 changes: 484 additions & 0 deletions demo/tutorials/tutorials_1.ipynb

Large diffs are not rendered by default.

3 changes: 2 additions & 1 deletion docker/GPU/Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -82,9 +82,10 @@ RUN cd /root/workspace/mmdeploy &&\
-DCMAKE_CXX_COMPILER=g++ \
-Dpplcv_DIR=/root/workspace/ppl.cv/cuda-build/install/lib/cmake/ppl \
-DTENSORRT_DIR=${TENSORRT_DIR} \
-DONNXRUNTIME_DIR=${ONNXRUNTIME_DIR} \
-DMMDEPLOY_BUILD_SDK_PYTHON_API=ON \
-DMMDEPLOY_TARGET_DEVICES="cuda;cpu" \
-DMMDEPLOY_TARGET_BACKENDS="trt" \
-DMMDEPLOY_TARGET_BACKENDS="ort;trt" \
-DMMDEPLOY_CODEBASES=all &&\
make -j$(nproc) && make install &&\
cd install/example && mkdir -p build && cd build &&\
Expand Down
4 changes: 2 additions & 2 deletions docs/en/01-how-to-build/android.md
Original file line number Diff line number Diff line change
Expand Up @@ -76,9 +76,9 @@ export OPENCV_ANDROID_SDK_DIR=${PWD}/OpenCV-android-sdk
<tr>
<td>ncnn </td>
<td>A high-performance neural network inference computing framework supporting for android.</br>
<b> Now, MMDeploy supports v20211208 and has to use <code>git clone</code> to download it.</b><br>
<b> Now, MMDeploy supports v20220216 and has to use <code>git clone</code> to download it.</b><br>
<pre><code>
git clone -b 20211208 https://github.com/Tencent/ncnn.git
git clone -b 20220216 https://github.com/Tencent/ncnn.git
cd ncnn
git submodule update --init
export NCNN_DIR=${PWD}
Expand Down
Loading