update test.sh for serving
This commit is contained in:
parent
bd025cd29b
commit
a2bd540c90
|
@ -77,9 +77,10 @@ fi
|
||||||
|
|
||||||
# prepare serving env
|
# prepare serving env
|
||||||
python_name=$(func_parser_value "${lines[2]}")
|
python_name=$(func_parser_value "${lines[2]}")
|
||||||
${python_name} -m pip install install paddle-serving-server-gpu==0.6.1.post101
|
wget https://paddle-serving.bj.bcebos.com/chain/paddle_serving_server_gpu-0.0.0.post101-py3-none-any.whl
|
||||||
|
${python_name} -m pip install install paddle_serving_server_gpu-0.0.0.post101-py3-none-any.whl
|
||||||
${python_name} -m pip install paddle_serving_client==0.6.1
|
${python_name} -m pip install paddle_serving_client==0.6.1
|
||||||
${python_name} -m pip install paddle-serving-app==0.6.1
|
${python_name} -m pip install paddle-serving-app==0.6.3
|
||||||
wget -nc -P ./inference https://paddleocr.bj.bcebos.com/dygraph_v2.0/ch/ch_ppocr_mobile_v2.0_det_infer.tar
|
wget -nc -P ./inference https://paddleocr.bj.bcebos.com/dygraph_v2.0/ch/ch_ppocr_mobile_v2.0_det_infer.tar
|
||||||
wget -nc -P ./inference https://paddleocr.bj.bcebos.com/dygraph_v2.0/ch/ch_ppocr_mobile_v2.0_rec_infer.tar
|
wget -nc -P ./inference https://paddleocr.bj.bcebos.com/dygraph_v2.0/ch/ch_ppocr_mobile_v2.0_rec_infer.tar
|
||||||
cd ./inference && tar xf ch_ppocr_mobile_v2.0_det_infer.tar && tar xf ch_ppocr_mobile_v2.0_rec_infer.tar
|
cd ./inference && tar xf ch_ppocr_mobile_v2.0_det_infer.tar && tar xf ch_ppocr_mobile_v2.0_rec_infer.tar
|
||||||
|
|
|
@ -168,7 +168,7 @@ web_use_trt_key=$(func_parser_key "${lines[63]}")
|
||||||
web_use_trt_list=$(func_parser_value "${lines[63]}")
|
web_use_trt_list=$(func_parser_value "${lines[63]}")
|
||||||
web_precision_key=$(func_parser_key "${lines[64]}")
|
web_precision_key=$(func_parser_key "${lines[64]}")
|
||||||
web_precision_list=$(func_parser_value "${lines[64]}")
|
web_precision_list=$(func_parser_value "${lines[64]}")
|
||||||
pipline_py=$(func_parser_value "${lines[65]}")
|
pipeline_py=$(func_parser_value "${lines[65]}")
|
||||||
|
|
||||||
|
|
||||||
LOG_PATH="./tests/output"
|
LOG_PATH="./tests/output"
|
||||||
|
@ -259,6 +259,8 @@ function func_serving(){
|
||||||
eval $trans_model_cmd
|
eval $trans_model_cmd
|
||||||
cd ${serving_dir_value}
|
cd ${serving_dir_value}
|
||||||
echo $PWD
|
echo $PWD
|
||||||
|
unset https_proxy
|
||||||
|
unset http_proxy
|
||||||
for use_gpu in ${web_use_gpu_list[*]}; do
|
for use_gpu in ${web_use_gpu_list[*]}; do
|
||||||
echo ${ues_gpu}
|
echo ${ues_gpu}
|
||||||
if [ ${use_gpu} = "null" ]; then
|
if [ ${use_gpu} = "null" ]; then
|
||||||
|
@ -269,12 +271,18 @@ function func_serving(){
|
||||||
for threads in ${web_cpu_threads_list[*]}; do
|
for threads in ${web_cpu_threads_list[*]}; do
|
||||||
_save_log_path="${_log_path}/server_cpu_usemkldnn_${use_mkldnn}_threads_${threads}_batchsize_1.log"
|
_save_log_path="${_log_path}/server_cpu_usemkldnn_${use_mkldnn}_threads_${threads}_batchsize_1.log"
|
||||||
set_cpu_threads=$(func_set_params "${web_cpu_threads_key}" "${threads}")
|
set_cpu_threads=$(func_set_params "${web_cpu_threads_key}" "${threads}")
|
||||||
web_service_cmd="${python} ${web_service_py} ${web_use_gpu_key}=${use_gpu} ${web_use_mkldnn_key}=${use_mkldnn} ${set_cpu_threads} &>log.txt &"
|
web_service_cmd="${python} ${web_service_py} ${web_use_gpu_key}=${use_gpu} ${web_use_mkldnn_key}=${use_mkldnn} ${set_cpu_threads} &>${_save_log_path} &"
|
||||||
echo $web_service_cmd
|
|
||||||
eval $web_service_cmd
|
eval $web_service_cmd
|
||||||
pipline_cmd="${python} ${pipline_py}"
|
sleep 2s
|
||||||
echo $pipline_cmd
|
pipeline_cmd="${python} ${pipeline_py}"
|
||||||
eval $pipline_cmd
|
eval $pipeline_cmd
|
||||||
|
last_status=${PIPESTATUS[0]}
|
||||||
|
eval "cat ${_save_log_path}"
|
||||||
|
status_check $last_status "${pipeline_cmd}" "${status_log}"
|
||||||
|
PID=$!
|
||||||
|
kill $PID
|
||||||
|
sleep 2s
|
||||||
|
ps ux | grep -E 'web_service|pipeline' | awk '{print $2}' | xargs kill -s 9
|
||||||
done
|
done
|
||||||
done
|
done
|
||||||
elif [ ${use_gpu} = "0" ]; then
|
elif [ ${use_gpu} = "0" ]; then
|
||||||
|
@ -286,18 +294,24 @@ function func_serving(){
|
||||||
if [[ ${precision} =~ "fp16" || ${precision} =~ "int8" ]] && [ ${use_trt} = "False" ]; then
|
if [[ ${precision} =~ "fp16" || ${precision} =~ "int8" ]] && [ ${use_trt} = "False" ]; then
|
||||||
continue
|
continue
|
||||||
fi
|
fi
|
||||||
if [[ ${use_trt} = "False" || ${precision} =~ "int8" ]] && [ ${_flag_quant} = "True" ]; then
|
if [[ ${use_trt} = "Falg_quantse" || ${precision} =~ "int8" ]]; then
|
||||||
continue
|
continue
|
||||||
fi
|
fi
|
||||||
_save_log_path="${_log_path}/infer_gpu_usetrt_${use_trt}_precision_${precision}_batchsize_1.log"
|
_save_log_path="${_log_path}/infer_gpu_usetrt_${use_trt}_precision_${precision}_batchsize_1.log"
|
||||||
set_tensorrt=$(func_set_params "${web_use_trt_key}" "${use_trt}")
|
set_tensorrt=$(func_set_params "${web_use_trt_key}" "${use_trt}")
|
||||||
set_precision=$(func_set_params "${web_precision_key}" "${precision}")
|
set_precision=$(func_set_params "${web_precision_key}" "${precision}")
|
||||||
web_service_cmd="${_python} ${web_service_py} ${web_use_gpu_key}=${use_gpu} ${web_use_trt_key}=${set_tensorrt} ${set_precision} > ${_save_log_path} &>log.txt & "
|
web_service_cmd="${python} ${web_service_py} ${web_use_gpu_key}=${use_gpu} ${set_tensorrt} ${set_precision} &>${_save_log_path} & "
|
||||||
echo $web_service_cmd
|
|
||||||
eval $web_service_cmd
|
eval $web_service_cmd
|
||||||
pipline_cmd="${python} ${pipline_py}"
|
sleep 2s
|
||||||
echo $pipline_cmd
|
pipeline_cmd="${python} ${pipeline_py}"
|
||||||
eval $pipline_cmd
|
eval $pipeline_cmd
|
||||||
|
last_status=${PIPESTATUS[0]}
|
||||||
|
eval "cat ${_save_log_path}"
|
||||||
|
status_check $last_status "${pipeline_cmd}" "${status_log}"
|
||||||
|
PID=$!
|
||||||
|
kill $PID
|
||||||
|
sleep 2s
|
||||||
|
ps ux | grep -E 'web_service|pipeline' | awk '{print $2}' | xargs kill -s 9
|
||||||
done
|
done
|
||||||
done
|
done
|
||||||
else
|
else
|
||||||
|
|
Loading…
Reference in New Issue