commit
dc6e724efb
|
@ -19,6 +19,7 @@ from __future__ import print_function
|
||||||
import paddle
|
import paddle
|
||||||
from paddle import ParamAttr
|
from paddle import ParamAttr
|
||||||
import paddle.nn as nn
|
import paddle.nn as nn
|
||||||
|
import paddle.nn.functional as F
|
||||||
|
|
||||||
__all__ = ["ResNet"]
|
__all__ = ["ResNet"]
|
||||||
|
|
||||||
|
@ -37,9 +38,9 @@ class ConvBNLayer(nn.Layer):
|
||||||
super(ConvBNLayer, self).__init__()
|
super(ConvBNLayer, self).__init__()
|
||||||
|
|
||||||
self.is_vd_mode = is_vd_mode
|
self.is_vd_mode = is_vd_mode
|
||||||
self._pool2d_avg = nn.AvgPool2d(
|
self._pool2d_avg = nn.AvgPool2D(
|
||||||
kernel_size=2, stride=2, padding=0, ceil_mode=True)
|
kernel_size=2, stride=2, padding=0, ceil_mode=True)
|
||||||
self._conv = nn.Conv2d(
|
self._conv = nn.Conv2D(
|
||||||
in_channels=in_channels,
|
in_channels=in_channels,
|
||||||
out_channels=out_channels,
|
out_channels=out_channels,
|
||||||
kernel_size=kernel_size,
|
kernel_size=kernel_size,
|
||||||
|
@ -118,7 +119,8 @@ class BottleneckBlock(nn.Layer):
|
||||||
short = inputs
|
short = inputs
|
||||||
else:
|
else:
|
||||||
short = self.short(inputs)
|
short = self.short(inputs)
|
||||||
y = paddle.elementwise_add(x=short, y=conv2, act='relu')
|
y = paddle.add(x=short, y=conv2)
|
||||||
|
y = F.relu(y)
|
||||||
return y
|
return y
|
||||||
|
|
||||||
|
|
||||||
|
@ -165,7 +167,8 @@ class BasicBlock(nn.Layer):
|
||||||
short = inputs
|
short = inputs
|
||||||
else:
|
else:
|
||||||
short = self.short(inputs)
|
short = self.short(inputs)
|
||||||
y = paddle.elementwise_add(x=short, y=conv1, act='relu')
|
y = paddle.add(x=short, y=conv1)
|
||||||
|
y = F.relu(y)
|
||||||
return y
|
return y
|
||||||
|
|
||||||
|
|
||||||
|
@ -214,7 +217,7 @@ class ResNet(nn.Layer):
|
||||||
stride=1,
|
stride=1,
|
||||||
act='relu',
|
act='relu',
|
||||||
name="conv1_3")
|
name="conv1_3")
|
||||||
self.pool2d_max = nn.MaxPool2d(kernel_size=3, stride=2, padding=1)
|
self.pool2d_max = nn.MaxPool2D(kernel_size=3, stride=2, padding=1)
|
||||||
|
|
||||||
self.stages = []
|
self.stages = []
|
||||||
self.out_channels = []
|
self.out_channels = []
|
||||||
|
|
|
@ -19,6 +19,7 @@ from __future__ import print_function
|
||||||
import paddle
|
import paddle
|
||||||
from paddle import ParamAttr
|
from paddle import ParamAttr
|
||||||
import paddle.nn as nn
|
import paddle.nn as nn
|
||||||
|
import paddle.nn.functional as F
|
||||||
|
|
||||||
__all__ = ["ResNet"]
|
__all__ = ["ResNet"]
|
||||||
|
|
||||||
|
@ -37,9 +38,9 @@ class ConvBNLayer(nn.Layer):
|
||||||
super(ConvBNLayer, self).__init__()
|
super(ConvBNLayer, self).__init__()
|
||||||
|
|
||||||
self.is_vd_mode = is_vd_mode
|
self.is_vd_mode = is_vd_mode
|
||||||
self._pool2d_avg = nn.AvgPool2d(
|
self._pool2d_avg = nn.AvgPool2D(
|
||||||
kernel_size=stride, stride=stride, padding=0, ceil_mode=True)
|
kernel_size=stride, stride=stride, padding=0, ceil_mode=True)
|
||||||
self._conv = nn.Conv2d(
|
self._conv = nn.Conv2D(
|
||||||
in_channels=in_channels,
|
in_channels=in_channels,
|
||||||
out_channels=out_channels,
|
out_channels=out_channels,
|
||||||
kernel_size=kernel_size,
|
kernel_size=kernel_size,
|
||||||
|
@ -119,7 +120,8 @@ class BottleneckBlock(nn.Layer):
|
||||||
short = inputs
|
short = inputs
|
||||||
else:
|
else:
|
||||||
short = self.short(inputs)
|
short = self.short(inputs)
|
||||||
y = paddle.elementwise_add(x=short, y=conv2, act='relu')
|
y = paddle.add(x=short, y=conv2)
|
||||||
|
y = F.relu(y)
|
||||||
return y
|
return y
|
||||||
|
|
||||||
|
|
||||||
|
@ -166,7 +168,8 @@ class BasicBlock(nn.Layer):
|
||||||
short = inputs
|
short = inputs
|
||||||
else:
|
else:
|
||||||
short = self.short(inputs)
|
short = self.short(inputs)
|
||||||
y = paddle.elementwise_add(x=short, y=conv1, act='relu')
|
y = paddle.add(x=short, y=conv1)
|
||||||
|
y = F.relu(y)
|
||||||
return y
|
return y
|
||||||
|
|
||||||
|
|
||||||
|
@ -215,7 +218,7 @@ class ResNet(nn.Layer):
|
||||||
stride=1,
|
stride=1,
|
||||||
act='relu',
|
act='relu',
|
||||||
name="conv1_3")
|
name="conv1_3")
|
||||||
self.pool2d_max = nn.MaxPool2d(kernel_size=3, stride=2, padding=1)
|
self.pool2d_max = nn.MaxPool2D(kernel_size=3, stride=2, padding=1)
|
||||||
|
|
||||||
self.block_list = []
|
self.block_list = []
|
||||||
if layers >= 50:
|
if layers >= 50:
|
||||||
|
@ -270,7 +273,7 @@ class ResNet(nn.Layer):
|
||||||
shortcut = True
|
shortcut = True
|
||||||
self.block_list.append(basic_block)
|
self.block_list.append(basic_block)
|
||||||
self.out_channels = num_filters[block]
|
self.out_channels = num_filters[block]
|
||||||
self.out_pool = nn.MaxPool2d(kernel_size=2, stride=2, padding=0)
|
self.out_pool = nn.MaxPool2D(kernel_size=2, stride=2, padding=0)
|
||||||
|
|
||||||
def forward(self, inputs):
|
def forward(self, inputs):
|
||||||
y = self.conv1_1(inputs)
|
y = self.conv1_1(inputs)
|
||||||
|
|
|
@ -18,6 +18,7 @@ from __future__ import print_function
|
||||||
|
|
||||||
import numpy as np
|
import numpy as np
|
||||||
import cv2
|
import cv2
|
||||||
|
import paddle
|
||||||
from shapely.geometry import Polygon
|
from shapely.geometry import Polygon
|
||||||
import pyclipper
|
import pyclipper
|
||||||
|
|
||||||
|
@ -130,7 +131,9 @@ class DBPostProcess(object):
|
||||||
return cv2.mean(bitmap[ymin:ymax + 1, xmin:xmax + 1], mask)[0]
|
return cv2.mean(bitmap[ymin:ymax + 1, xmin:xmax + 1], mask)[0]
|
||||||
|
|
||||||
def __call__(self, pred, shape_list):
|
def __call__(self, pred, shape_list):
|
||||||
pred = pred.numpy()[:, 0, :, :]
|
if isinstance(pred, paddle.Tensor):
|
||||||
|
pred = pred.numpy()
|
||||||
|
pred = pred[:, 0, :, :]
|
||||||
segmentation = pred > self.thresh
|
segmentation = pred > self.thresh
|
||||||
|
|
||||||
boxes_batch = []
|
boxes_batch = []
|
||||||
|
@ -140,4 +143,4 @@ class DBPostProcess(object):
|
||||||
pred[batch_index], segmentation[batch_index], width, height)
|
pred[batch_index], segmentation[batch_index], width, height)
|
||||||
|
|
||||||
boxes_batch.append({'points': boxes})
|
boxes_batch.append({'points': boxes})
|
||||||
return boxes_batch
|
return boxes_batch
|
|
@ -1,4 +1,5 @@
|
||||||
import cv2
|
import cv2
|
||||||
|
import paddle
|
||||||
import numpy as np
|
import numpy as np
|
||||||
import pyclipper
|
import pyclipper
|
||||||
from shapely.geometry import Polygon
|
from shapely.geometry import Polygon
|
||||||
|
@ -23,7 +24,9 @@ class DBPostProcess():
|
||||||
pred:
|
pred:
|
||||||
binary: text region segmentation map, with shape (N, 1,H, W)
|
binary: text region segmentation map, with shape (N, 1,H, W)
|
||||||
'''
|
'''
|
||||||
pred = pred.numpy()[:, 0, :, :]
|
if isinstance(pred, paddle.Tensor):
|
||||||
|
pred = pred.numpy()
|
||||||
|
pred = pred[:, 0, :, :]
|
||||||
segmentation = self.binarize(pred)
|
segmentation = self.binarize(pred)
|
||||||
batch_out = []
|
batch_out = []
|
||||||
for batch_index in range(pred.shape[0]):
|
for batch_index in range(pred.shape[0]):
|
||||||
|
@ -130,4 +133,4 @@ class DBPostProcess():
|
||||||
box[:, 0] = box[:, 0] - xmin
|
box[:, 0] = box[:, 0] - xmin
|
||||||
box[:, 1] = box[:, 1] - ymin
|
box[:, 1] = box[:, 1] - ymin
|
||||||
cv2.fillPoly(mask, box.reshape(1, -1, 2).astype(np.int32), 1)
|
cv2.fillPoly(mask, box.reshape(1, -1, 2).astype(np.int32), 1)
|
||||||
return cv2.mean(bitmap[ymin:ymax + 1, xmin:xmax + 1], mask)[0]
|
return cv2.mean(bitmap[ymin:ymax + 1, xmin:xmax + 1], mask)[0]
|
|
@ -100,9 +100,10 @@ class CTCLabelDecode(BaseRecLabelDecode):
|
||||||
character_type, use_space_char)
|
character_type, use_space_char)
|
||||||
|
|
||||||
def __call__(self, preds, label=None, *args, **kwargs):
|
def __call__(self, preds, label=None, *args, **kwargs):
|
||||||
|
if isinstance(preds, paddle.Tensor):
|
||||||
|
preds = preds.numpy()
|
||||||
# out = self.decode_preds(preds)
|
# out = self.decode_preds(preds)
|
||||||
|
|
||||||
preds = F.softmax(preds, axis=2).numpy()
|
|
||||||
preds_idx = preds.argmax(axis=2)
|
preds_idx = preds.argmax(axis=2)
|
||||||
preds_prob = preds.max(axis=2)
|
preds_prob = preds.max(axis=2)
|
||||||
text = self.decode(preds_idx, preds_prob)
|
text = self.decode(preds_idx, preds_prob)
|
||||||
|
@ -116,19 +117,18 @@ class CTCLabelDecode(BaseRecLabelDecode):
|
||||||
return dict_character
|
return dict_character
|
||||||
|
|
||||||
def decode_preds(self, preds):
|
def decode_preds(self, preds):
|
||||||
probs = F.softmax(preds, axis=2).numpy()
|
probs_ind = np.argmax(preds, axis=2)
|
||||||
probs_ind = np.argmax(probs, axis=2)
|
|
||||||
|
|
||||||
B, N, _ = preds.shape
|
B, N, _ = preds.shape
|
||||||
l = np.ones(B).astype(np.int64) * N
|
l = np.ones(B).astype(np.int64) * N
|
||||||
length = paddle.to_variable(l)
|
length = paddle.to_tensor(l)
|
||||||
out = paddle.fluid.layers.ctc_greedy_decoder(preds, 0, length)
|
out = paddle.fluid.layers.ctc_greedy_decoder(preds, 0, length)
|
||||||
batch_res = [
|
batch_res = [
|
||||||
x[:idx[0]] for x, idx in zip(out[0].numpy(), out[1].numpy())
|
x[:idx[0]] for x, idx in zip(out[0].numpy(), out[1].numpy())
|
||||||
]
|
]
|
||||||
|
|
||||||
result_list = []
|
result_list = []
|
||||||
for sample_idx, ind, prob in zip(batch_res, probs_ind, probs):
|
for sample_idx, ind, prob in zip(batch_res, probs_ind, preds):
|
||||||
char_list = [self.character[idx] for idx in sample_idx]
|
char_list = [self.character[idx] for idx in sample_idx]
|
||||||
valid_ind = np.where(ind != 0)[0]
|
valid_ind = np.where(ind != 0)[0]
|
||||||
if len(valid_ind) == 0:
|
if len(valid_ind) == 0:
|
||||||
|
@ -172,4 +172,4 @@ class AttnLabelDecode(BaseRecLabelDecode):
|
||||||
else:
|
else:
|
||||||
assert False, "unsupport type %s in get_beg_end_flag_idx" \
|
assert False, "unsupport type %s in get_beg_end_flag_idx" \
|
||||||
% beg_or_end
|
% beg_or_end
|
||||||
return idx
|
return idx
|
|
@ -68,11 +68,11 @@ def load_dygraph_pretrain(model, logger, path=None, load_static_weights=False):
|
||||||
param_state_dict[key] = pre_state_dict[weight_name]
|
param_state_dict[key] = pre_state_dict[weight_name]
|
||||||
else:
|
else:
|
||||||
param_state_dict[key] = model_dict[key]
|
param_state_dict[key] = model_dict[key]
|
||||||
model.set_dict(param_state_dict)
|
model.set_state_dict(param_state_dict)
|
||||||
return
|
return
|
||||||
|
|
||||||
param_state_dict, optim_state_dict = paddle.load(path)
|
param_state_dict = paddle.load(path + '.pdparams')
|
||||||
model.set_dict(param_state_dict)
|
model.set_state_dict(param_state_dict)
|
||||||
return
|
return
|
||||||
|
|
||||||
|
|
||||||
|
@ -91,7 +91,7 @@ def init_model(config, model, logger, optimizer=None, lr_scheduler=None):
|
||||||
"Given dir {}.pdopt not exist.".format(checkpoints)
|
"Given dir {}.pdopt not exist.".format(checkpoints)
|
||||||
para_dict = paddle.load(checkpoints + '.pdparams')
|
para_dict = paddle.load(checkpoints + '.pdparams')
|
||||||
opti_dict = paddle.load(checkpoints + '.pdopt')
|
opti_dict = paddle.load(checkpoints + '.pdopt')
|
||||||
model.set_dict(para_dict)
|
model.set_state_dict(para_dict)
|
||||||
if optimizer is not None:
|
if optimizer is not None:
|
||||||
optimizer.set_state_dict(opti_dict)
|
optimizer.set_state_dict(opti_dict)
|
||||||
|
|
||||||
|
|
|
@ -12,6 +12,13 @@
|
||||||
# See the License for the specific language governing permissions and
|
# See the License for the specific language governing permissions and
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
|
|
||||||
|
import os
|
||||||
|
import sys
|
||||||
|
|
||||||
|
__dir__ = os.path.dirname(os.path.abspath(__file__))
|
||||||
|
sys.path.append(__dir__)
|
||||||
|
sys.path.append(os.path.abspath(os.path.join(__dir__, '..')))
|
||||||
|
|
||||||
import argparse
|
import argparse
|
||||||
|
|
||||||
import paddle
|
import paddle
|
||||||
|
@ -20,14 +27,11 @@ from paddle.jit import to_static
|
||||||
from ppocr.modeling.architectures import build_model
|
from ppocr.modeling.architectures import build_model
|
||||||
from ppocr.postprocess import build_post_process
|
from ppocr.postprocess import build_post_process
|
||||||
from ppocr.utils.save_load import init_model
|
from ppocr.utils.save_load import init_model
|
||||||
|
from ppocr.utils.logging import get_logger
|
||||||
from tools.program import load_config
|
from tools.program import load_config
|
||||||
from tools.program import merge_config
|
|
||||||
|
|
||||||
|
|
||||||
def parse_args():
|
def parse_args():
|
||||||
def str2bool(v):
|
|
||||||
return v.lower() in ("true", "t", "1")
|
|
||||||
|
|
||||||
parser = argparse.ArgumentParser()
|
parser = argparse.ArgumentParser()
|
||||||
parser.add_argument("-c", "--config", help="configuration file to use")
|
parser.add_argument("-c", "--config", help="configuration file to use")
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
|
@ -43,7 +47,7 @@ class Model(paddle.nn.Layer):
|
||||||
# Please modify the 'shape' according to actual needs
|
# Please modify the 'shape' according to actual needs
|
||||||
@to_static(input_spec=[
|
@to_static(input_spec=[
|
||||||
paddle.static.InputSpec(
|
paddle.static.InputSpec(
|
||||||
shape=[None, 3, 32, None], dtype='float32')
|
shape=[None, 3, 640, 640], dtype='float32')
|
||||||
])
|
])
|
||||||
def forward(self, inputs):
|
def forward(self, inputs):
|
||||||
x = self.pre_model(inputs)
|
x = self.pre_model(inputs)
|
||||||
|
@ -53,14 +57,13 @@ class Model(paddle.nn.Layer):
|
||||||
def main():
|
def main():
|
||||||
FLAGS = parse_args()
|
FLAGS = parse_args()
|
||||||
config = load_config(FLAGS.config)
|
config = load_config(FLAGS.config)
|
||||||
merge_config(FLAGS.opt)
|
logger = get_logger()
|
||||||
|
|
||||||
# build post process
|
# build post process
|
||||||
post_process_class = build_post_process(config['PostProcess'],
|
post_process_class = build_post_process(config['PostProcess'],
|
||||||
config['Global'])
|
config['Global'])
|
||||||
|
|
||||||
# build model
|
# build model
|
||||||
#for rec algorithm
|
# for rec algorithm
|
||||||
if hasattr(post_process_class, 'character'):
|
if hasattr(post_process_class, 'character'):
|
||||||
char_num = len(getattr(post_process_class, 'character'))
|
char_num = len(getattr(post_process_class, 'character'))
|
||||||
config['Architecture']["Head"]['out_channels'] = char_num
|
config['Architecture']["Head"]['out_channels'] = char_num
|
||||||
|
@ -69,7 +72,10 @@ def main():
|
||||||
model.eval()
|
model.eval()
|
||||||
|
|
||||||
model = Model(model)
|
model = Model(model)
|
||||||
paddle.jit.save(model, FLAGS.output_path)
|
save_path = '{}/{}'.format(FLAGS.output_path,
|
||||||
|
config['Architecture']['model_type'])
|
||||||
|
paddle.jit.save(model, save_path)
|
||||||
|
logger.info('inference model is saved to {}'.format(save_path))
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
|
|
|
@ -22,7 +22,6 @@ import cv2
|
||||||
import numpy as np
|
import numpy as np
|
||||||
import time
|
import time
|
||||||
import sys
|
import sys
|
||||||
|
|
||||||
import paddle
|
import paddle
|
||||||
|
|
||||||
import tools.infer.utility as utility
|
import tools.infer.utility as utility
|
||||||
|
@ -39,7 +38,7 @@ class TextDetector(object):
|
||||||
postprocess_params = {}
|
postprocess_params = {}
|
||||||
if self.det_algorithm == "DB":
|
if self.det_algorithm == "DB":
|
||||||
pre_process_list = [{
|
pre_process_list = [{
|
||||||
'ResizeForTest': {
|
'DetResizeForTest': {
|
||||||
'limit_side_len': args.det_limit_side_len,
|
'limit_side_len': args.det_limit_side_len,
|
||||||
'limit_type': args.det_limit_type
|
'limit_type': args.det_limit_type
|
||||||
}
|
}
|
||||||
|
@ -53,7 +52,7 @@ class TextDetector(object):
|
||||||
}, {
|
}, {
|
||||||
'ToCHWImage': None
|
'ToCHWImage': None
|
||||||
}, {
|
}, {
|
||||||
'keepKeys': {
|
'KeepKeys': {
|
||||||
'keep_keys': ['image', 'shape']
|
'keep_keys': ['image', 'shape']
|
||||||
}
|
}
|
||||||
}]
|
}]
|
||||||
|
@ -68,8 +67,9 @@ class TextDetector(object):
|
||||||
|
|
||||||
self.preprocess_op = create_operators(pre_process_list)
|
self.preprocess_op = create_operators(pre_process_list)
|
||||||
self.postprocess_op = build_post_process(postprocess_params)
|
self.postprocess_op = build_post_process(postprocess_params)
|
||||||
self.predictor = paddle.jit.load(args.det_model_dir)
|
self.predictor, self.input_tensor, self.output_tensors = utility.create_predictor(
|
||||||
self.predictor.eval()
|
args, 'det', logger) # paddle.jit.load(args.det_model_dir)
|
||||||
|
# self.predictor.eval()
|
||||||
|
|
||||||
def order_points_clockwise(self, pts):
|
def order_points_clockwise(self, pts):
|
||||||
"""
|
"""
|
||||||
|
@ -133,11 +133,23 @@ class TextDetector(object):
|
||||||
return None, 0
|
return None, 0
|
||||||
img = np.expand_dims(img, axis=0)
|
img = np.expand_dims(img, axis=0)
|
||||||
shape_list = np.expand_dims(shape_list, axis=0)
|
shape_list = np.expand_dims(shape_list, axis=0)
|
||||||
|
img = img.copy()
|
||||||
starttime = time.time()
|
starttime = time.time()
|
||||||
|
|
||||||
preds = self.predictor(img)
|
if self.use_zero_copy_run:
|
||||||
post_result = self.postprocess_op(preds, shape_list)
|
self.input_tensor.copy_from_cpu(img)
|
||||||
|
self.predictor.zero_copy_run()
|
||||||
|
else:
|
||||||
|
im = paddle.fluid.core.PaddleTensor(img)
|
||||||
|
self.predictor.run([im])
|
||||||
|
outputs = []
|
||||||
|
for output_tensor in self.output_tensors:
|
||||||
|
output = output_tensor.copy_to_cpu()
|
||||||
|
outputs.append(output)
|
||||||
|
preds = outputs[0]
|
||||||
|
|
||||||
|
# preds = self.predictor(img)
|
||||||
|
post_result = self.postprocess_op(preds, shape_list)
|
||||||
dt_boxes = post_result[0]['points']
|
dt_boxes = post_result[0]['points']
|
||||||
dt_boxes = self.filter_tag_det_res(dt_boxes, ori_im.shape)
|
dt_boxes = self.filter_tag_det_res(dt_boxes, ori_im.shape)
|
||||||
elapse = time.time() - starttime
|
elapse = time.time() - starttime
|
||||||
|
@ -146,8 +158,6 @@ class TextDetector(object):
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
args = utility.parse_args()
|
args = utility.parse_args()
|
||||||
place = paddle.CPUPlace()
|
|
||||||
paddle.disable_static(place)
|
|
||||||
|
|
||||||
image_file_list = get_image_file_list(args.image_dir)
|
image_file_list = get_image_file_list(args.image_dir)
|
||||||
logger = get_logger()
|
logger = get_logger()
|
||||||
|
|
|
@ -29,12 +29,11 @@ import cv2
|
||||||
import json
|
import json
|
||||||
import paddle
|
import paddle
|
||||||
|
|
||||||
from ppocr.utils.logging import get_logger
|
|
||||||
from ppocr.data import create_operators, transform
|
from ppocr.data import create_operators, transform
|
||||||
from ppocr.modeling import build_model
|
from ppocr.modeling.architectures import build_model
|
||||||
from ppocr.postprocess import build_post_process
|
from ppocr.postprocess import build_post_process
|
||||||
from ppocr.utils.save_load import init_model
|
from ppocr.utils.save_load import init_model
|
||||||
from ppocr.utils.utility import print_dict, get_image_file_list
|
from ppocr.utils.utility import get_image_file_list
|
||||||
import tools.program as program
|
import tools.program as program
|
||||||
|
|
||||||
|
|
||||||
|
@ -67,11 +66,11 @@ def main():
|
||||||
|
|
||||||
# create data ops
|
# create data ops
|
||||||
transforms = []
|
transforms = []
|
||||||
for op in config['EVAL']['dataset']['transforms']:
|
for op in config['Eval']['dataset']['transforms']:
|
||||||
op_name = list(op)[0]
|
op_name = list(op)[0]
|
||||||
if 'Label' in op_name:
|
if 'Label' in op_name:
|
||||||
continue
|
continue
|
||||||
elif op_name == 'keepKeys':
|
elif op_name == 'KeepKeys':
|
||||||
op[op_name]['keep_keys'] = ['image', 'shape']
|
op[op_name]['keep_keys'] = ['image', 'shape']
|
||||||
transforms.append(op)
|
transforms.append(op)
|
||||||
|
|
||||||
|
@ -92,8 +91,7 @@ def main():
|
||||||
|
|
||||||
images = np.expand_dims(batch[0], axis=0)
|
images = np.expand_dims(batch[0], axis=0)
|
||||||
shape_list = np.expand_dims(batch[1], axis=0)
|
shape_list = np.expand_dims(batch[1], axis=0)
|
||||||
images = paddle.to_variable(images)
|
images = paddle.to_tensor(images)
|
||||||
print(images.shape)
|
|
||||||
preds = model(images)
|
preds = model(images)
|
||||||
post_result = post_process_class(preds, shape_list)
|
post_result = post_process_class(preds, shape_list)
|
||||||
boxes = post_result[0]['points']
|
boxes = post_result[0]['points']
|
||||||
|
@ -109,14 +107,7 @@ def main():
|
||||||
draw_det_res(boxes, config, src_img, file)
|
draw_det_res(boxes, config, src_img, file)
|
||||||
logger.info("success!")
|
logger.info("success!")
|
||||||
|
|
||||||
# save inference model
|
|
||||||
# paddle.jit.save(model, 'output/model')
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
place, config = program.preprocess()
|
config, device, logger, vdl_writer = program.preprocess()
|
||||||
paddle.disable_static(place)
|
main()
|
||||||
|
|
||||||
logger = get_logger()
|
|
||||||
print_dict(config, logger)
|
|
||||||
main()
|
|
|
@ -27,12 +27,11 @@ sys.path.append(os.path.abspath(os.path.join(__dir__, '..')))
|
||||||
|
|
||||||
import paddle
|
import paddle
|
||||||
|
|
||||||
from ppocr.utils.logging import get_logger
|
|
||||||
from ppocr.data import create_operators, transform
|
from ppocr.data import create_operators, transform
|
||||||
from ppocr.modeling import build_model
|
from ppocr.modeling.architectures import build_model
|
||||||
from ppocr.postprocess import build_post_process
|
from ppocr.postprocess import build_post_process
|
||||||
from ppocr.utils.save_load import init_model
|
from ppocr.utils.save_load import init_model
|
||||||
from ppocr.utils.utility import print_dict, get_image_file_list
|
from ppocr.utils.utility import get_image_file_list
|
||||||
import tools.program as program
|
import tools.program as program
|
||||||
|
|
||||||
|
|
||||||
|
@ -54,13 +53,13 @@ def main():
|
||||||
|
|
||||||
# create data ops
|
# create data ops
|
||||||
transforms = []
|
transforms = []
|
||||||
for op in config['EVAL']['dataset']['transforms']:
|
for op in config['Eval']['dataset']['transforms']:
|
||||||
op_name = list(op)[0]
|
op_name = list(op)[0]
|
||||||
if 'Label' in op_name:
|
if 'Label' in op_name:
|
||||||
continue
|
continue
|
||||||
elif op_name in ['RecResizeImg']:
|
elif op_name in ['RecResizeImg']:
|
||||||
op[op_name]['infer_mode'] = True
|
op[op_name]['infer_mode'] = True
|
||||||
elif op_name == 'keepKeys':
|
elif op_name == 'KeepKeys':
|
||||||
op[op_name]['keep_keys'] = ['image']
|
op[op_name]['keep_keys'] = ['image']
|
||||||
transforms.append(op)
|
transforms.append(op)
|
||||||
global_config['infer_mode'] = True
|
global_config['infer_mode'] = True
|
||||||
|
@ -75,22 +74,14 @@ def main():
|
||||||
batch = transform(data, ops)
|
batch = transform(data, ops)
|
||||||
|
|
||||||
images = np.expand_dims(batch[0], axis=0)
|
images = np.expand_dims(batch[0], axis=0)
|
||||||
images = paddle.to_variable(images)
|
images = paddle.to_tensor(images)
|
||||||
preds = model(images)
|
preds = model(images)
|
||||||
post_result = post_process_class(preds)
|
post_result = post_process_class(preds)
|
||||||
for rec_reuslt in post_result:
|
for rec_reuslt in post_result:
|
||||||
logger.info('\t result: {}'.format(rec_reuslt))
|
logger.info('\t result: {}'.format(rec_reuslt))
|
||||||
logger.info("success!")
|
logger.info("success!")
|
||||||
|
|
||||||
# save inference model
|
|
||||||
# currently, paddle.jit.to_static not support rnn
|
|
||||||
# paddle.jit.save(model, 'output/rec/model')
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
place, config = program.preprocess()
|
config, device, logger, vdl_writer = program.preprocess()
|
||||||
paddle.disable_static(place)
|
|
||||||
|
|
||||||
logger = get_logger()
|
|
||||||
print_dict(config, logger)
|
|
||||||
main()
|
main()
|
||||||
|
|
Loading…
Reference in New Issue