Skip to content

Latest commit

 

History

History
 
 

Folders and files

NameName
Last commit message
Last commit date

parent directory

..
 
 
 
 
 
 
 
 

English | 简体中文

PP-Matting C++ Deployment Example

This directory provides examples that infer.cc fast finishes the deployment of PP-Matting on CPU/GPU and GPU accelerated by TensorRT. Before deployment, two steps require confirmation

Taking the PP-Matting inference on Linux as an example, the compilation test can be completed by executing the following command in this directory. FastDeploy version 0.7.0 or above (x.x.x>=0.7.0) is required to support this model.

mkdir build
cd build
# Download the FastDeploy precompiled library. Users can choose your appropriate version in the `FastDeploy  Precompiled Library` mentioned above 
wget https://bj.bcebos.com/fastdeploy/release/cpp/fastdeploy-linux-x64-x.x.x.tgz
tar xvf fastdeploy-linux-x64-x.x.x.tgz
cmake .. -DFASTDEPLOY_INSTALL_DIR=${PWD}/fastdeploy-linux-x64-x.x.x
make -j

# Download PP-Matting model files and test images
wget https://bj.bcebos.com/paddlehub/fastdeploy/PP-Matting-512.tgz
tar -xvf PP-Matting-512.tgz
wget https://bj.bcebos.com/paddlehub/fastdeploy/matting_input.jpg
wget https://bj.bcebos.com/paddlehub/fastdeploy/matting_bgr.jpg


# CPU inference
./infer_demo PP-Matting-512 matting_input.jpg matting_bgr.jpg 0
# GPU inference
./infer_demo PP-Matting-512 matting_input.jpg matting_bgr.jpg 1
# TensorRT inference on GPU
./infer_demo PP-Matting-512 matting_input.jpg matting_bgr.jpg 2
# kunlunxin XPU inference
./infer_demo PP-Matting-512 matting_input.jpg matting_bgr.jpg 3

The visualized result after running is as follows

The above command works for Linux or MacOS. For SDK use-pattern in Windows, refer to:

PP-Matting C++ Interface

PPMatting Class

fastdeploy::vision::matting::PPMatting(
        const string& model_file,
        const string& params_file = "",
        const string& config_file,
        const RuntimeOption& runtime_option = RuntimeOption(),
        const ModelFormat& model_format = ModelFormat::PADDLE)

PP-Matting model loading and initialization, among which model_file is the exported Paddle model format.

Parameter

  • model_file(str): Model file path
  • params_file(str): Parameter file path
  • config_file(str): Inference deployment configuration file
  • runtime_option(RuntimeOption): Backend inference configuration. None by default, which is the default configuration
  • model_format(ModelFormat): Model format. Paddle format by default

Predict Function

PPMatting::Predict(cv::Mat* im, MattingResult* result)

Model prediction interface. Input images and output detection results.

Parameter

  • im: Input images in HWC or BGR format
  • result: The segmentation result, including the predicted label of the segmentation and the corresponding probability of the label. Refer to Vision Model Prediction Results for the description of SegmentationResult

Class Member Variable

Pre-processing Parameter

Users can modify the following pre-processing parameters to their needs, which affects the final inference and deployment results