2023-03-03 14:51:09 +08:00
|
|
|
|
# optional, If name is specified it must match the name of the model repository directory containing the model.
|
|
|
|
|
|
name: "rec_runtime"
|
|
|
|
|
|
backend: "fastdeploy"
|
|
|
|
|
|
max_batch_size: 128
|
|
|
|
|
|
|
|
|
|
|
|
# Input configuration of the model
|
|
|
|
|
|
input [
|
|
|
|
|
|
{
|
|
|
|
|
|
# input name
|
|
|
|
|
|
name: "x"
|
|
|
|
|
|
# input type such as TYPE_FP32、TYPE_UINT8、TYPE_INT8、TYPE_INT16、TYPE_INT32、TYPE_INT64、TYPE_FP16、TYPE_STRING
|
|
|
|
|
|
data_type: TYPE_FP32
|
|
|
|
|
|
# input shape, The batch dimension is omitted and the actual shape is [batch, c, h, w]
|
|
|
|
|
|
dims: [ 3, 48, -1 ]
|
|
|
|
|
|
}
|
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
|
|
# The output of the model is configured in the same format as the input
|
|
|
|
|
|
output [
|
|
|
|
|
|
{
|
|
|
|
|
|
name: "softmax_5.tmp_0"
|
|
|
|
|
|
data_type: TYPE_FP32
|
|
|
|
|
|
dims: [ -1, 6625 ]
|
|
|
|
|
|
}
|
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
|
|
# Number of instances of the model
|
|
|
|
|
|
instance_group [
|
|
|
|
|
|
{
|
|
|
|
|
|
# The number of instances is 1
|
|
|
|
|
|
count: 1
|
|
|
|
|
|
# Use GPU, CPU inference option is:KIND_CPU
|
|
|
|
|
|
kind: KIND_GPU
|
|
|
|
|
|
# The instance is deployed on the 0th GPU card
|
|
|
|
|
|
gpus: [0]
|
|
|
|
|
|
}
|
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
|
|
optimization {
|
|
|
|
|
|
execution_accelerators {
|
|
|
|
|
|
# GPU推理配置, 配合KIND_GPU使用
|
|
|
|
|
|
gpu_execution_accelerator : [
|
|
|
|
|
|
{
|
|
|
|
|
|
name : "paddle"
|
|
|
|
|
|
# 设置推理并行计算线程数为4
|
|
|
|
|
|
parameters { key: "cpu_threads" value: "4" }
|
|
|
|
|
|
# 开启mkldnn加速,设置为0关闭mkldnn
|
|
|
|
|
|
parameters { key: "use_mkldnn" value: "1" }
|
|
|
|
|
|
}
|
|
|
|
|
|
]
|
|
|
|
|
|
}
|
2024-05-29 15:26:09 +08:00
|
|
|
|
}
|