如何在Keras模型内部为新的仅推理的端到端模型进行预处理



使用Tensorflow 2.3,我已经用EfficientNet训练了一个tf.keras模型,现在我想导出一个仅推理的端到端模型,该模型将包括预处理层(本质上是解码编码为base64字符串的图像,并可能处理归一化(。我定义训练模型和推理模型的代码:

imageSize = (224,224,3)
inputs = layers.Input(shape=imageSize)
eff_net_model = EfficientNetB0(input_tensor=inputs)
eff_net_model.compile(...)
eff_net_model.fit(...)
# training is finished now wrap the model with preprocessing for inference model
model = b64_image_model_wrapper(eff_net_model,imageSize)

我的包装fct如下:

def b64_image_model_wrapper(model,imageSize,method=tf.image.ResizeMethod.BILINEAR, 
mean=0.0,std=1.0,input_name="b64_image"):
def preprocess_and_decode(img_str, new_shape=imageSize):
img = tf.io.decode_base64(img_str)
img = tf.image.decode_jpeg(img, channels=3)
img = tf.image.resize(img, new_shape, method=method)
return img
input64 = tf.keras.layers.Input(shape=(None,), dtype="string", name=input_name)
output_tensor = tf.keras.layers.Lambda(
lambda img: tf.map_fn(lambda im: preprocess_and_decode(im[0]), img, fn_output_signature=tf.float32,swap_memory=True))(input64)
x = (output_tensor - mean) / std
x = model(x)
new_model = tf.keras.Model(input64, x, name=model.name)
new_model.output_names = model.output_names
return new_model

稍后,如果我想用这个新模型(将字符串编码的图像作为输入的模型(进行预测,我会这样做:

model.predict(np.array(["hello"])) # just for testing purposes

我得到错误:

Exception has occurred: InvalidArgumentError
Expected image (JPEG, PNG, or GIF), got unknown format starting with '261330_201250354205327340i327'
[[{{node EfficientNet/lambda/map/while/body/_1/EfficientNet/lambda/map/while/DecodeJpeg}}]]

同样,如果我保存这个新模型(tf.saved_model.save(model)(并尝试将其与Tensorflow Serving一起使用,我会得到错误:

Error while reading resource variable block6c_bn/moving_mean from Container: localhost. This could mean that the variable was uninitialized. Not found: Container localhost does not exist. (Could not find resource: localhost/block6c_bn/moving_mean)nt [[{{node functional_1/EfficientNet/block6c_bn/FusedBatchNormV3/ReadVariableOp}}]]

我不知道这到底意味着什么,但似乎有些权重没有初始化?我做模型包装的方式对吗?

调用model.predict(np.array(["hello"]))时出错的原因是:

  1. 输入格式不是base64字符串
  2. tf.io.decode_base64之后的输入不是.jpg文件

以下代码向您展示如何使用.jpg文件测试模型:

with open("./homersimpson.0.0.jpg", "rb") as imageFile:
base64_bytes = base64.urlsafe_b64encode(imageFile.read())
base64_strings = base64_bytes.decode('ascii')
img = tf.io.decode_base64(base64_strings)
prediction = model.predict(np.array([base64_strings]))

此外,我使用model.save('./test_save_model')来保存整个模型,它没有任何问题

最新更新