用于多输入图像的 VGG16 网络

2024-04-28

我正在尝试将 VGG16 网络用于多个输入图像。 使用具有 2 个输入的简单 CNN 训练该模型给了我一个 acc。大约 50%,这就是为什么我想使用 VGG16 这样的既定模型进行尝试。
这是我尝试过的:

# imports
from keras.applications.vgg16 import VGG16
from keras.models import Model
from keras.layers import Conv2D, MaxPooling2D, Activation, Dropout, Flatten, Dense

def def_model():
    model = VGG16(include_top=False, input_shape=(224, 224, 3))
    # mark loaded layers as not trainable
    for layer in model.layers:
        layer.trainable = False
    # return last pooling layer
    pool_layer = model.layers[-1].output
    return pool_layer

m1 = def_model()
m2 = def_model() 
m3 = def_model()

# add classifier layers
merge = concatenate([m1, m2, m3])

# optinal_conv = Conv2D(64, (3, 3), activation='relu', padding='same')(merge)
# optinal_pool = MaxPooling2D(pool_size=(2, 2))(optinal_conv)
# flatten = Flatten()(optinal_pool)

flatten = Flatten()(merge)
dense1 = Dense(512, activation='relu')(flatten)
dense2 = Dropout(0.5)(dense1)
output = Dense(1, activation='sigmoid')(dense2)


inshape1 = Input(shape=(224, 224, 3))
inshape2 = Input(shape=(224, 224, 3))
inshape3 = Input(shape=(224, 224, 3))
model = Model(inputs=[inshape1, inshape2, inshape3], outputs=output)

  1. 我在调用时收到此错误Model功能。
ValueError: Graph disconnected: cannot obtain value for tensor Tensor("input_21:0", shape=(?, 224, 224, 3), dtype=float32) at layer "input_21". The following previous layers were accessed without issue: []`

我知道该图是断开的,但我找不到哪里。
这里有compile and fit功能。

# compile model
model.compile(optimizer="Adam", loss='binary_crossentropy', metrics=['accuracy'])
model.fit([train1, train2, train3], train, 
           validation_data=([test1, test2, test3], ytest))
  1. 我对某些行发表了评论:optinal_conv and optinal_pool。应用后可能有什么效果Conv2D and MaxPooling2D之后concatenate功能?

我建议看看这个答案具有 Keras 功能 API 的多输入多输出模型 https://stackoverflow.com/a/66849164/9215780。这是实现这一目标的一种方法:

# 3 inputs 
input0 = tf.keras.Input(shape=(224, 224, 3), name="img0")
input1 = tf.keras.Input(shape=(224, 224, 3), name="img1")
input2 = tf.keras.Input(shape=(224, 224, 3), name="img2")
concate_input = tf.keras.layers.Concatenate()([input0, input1, input2])
# get 3 feature maps with same size (224, 224)
# pretrained models needs that
input = tf.keras.layers.Conv2D(3, (3, 3), 
                     padding='same', activation="relu")(concate_input)

# pass that to imagenet model 
vg = tf.keras.applications.VGG16(weights=None,
                                 include_top = False, 
                                 input_tensor = input)

# do whatever 
gap = tf.keras.layers.GlobalAveragePooling2D()(vg.output)
den = tf.keras.layers.Dense(1, activation='sigmoid')(gap)

# build the complete model 
model = tf.keras.Model(inputs=[input0, input1, input2], outputs=den)
本文内容由网友自发贡献,版权归原作者所有,本站不承担相应法律责任。如您发现有涉嫌抄袭侵权的内容,请联系:hwhale#tublm.com(使用前将#替换为@)

用于多输入图像的 VGG16 网络 的相关文章

随机推荐