android 缓冲区大小错误的 Tensorflowlite

Tensorflowlite on android buffer size error

我正在尝试构建图像分类器 android 应用程序。我已经使用 keras 构建了我的模型。 型号如下:

model.add(MobileNetV2(include_top=False, weights='imagenet',input_shape=(224, 224, 3)))
model.add(GlobalAveragePooling2D())
model.add(Dropout(0.5))
model.add(Dense(3, activation='softmax'))

model.layers[0].trainable = False     
model.compile(optimizer='adam',  loss='categorical_crossentropy', metrics=['accuracy'])
model.summary()

输出:

Model: "sequential_3"
_________________________________________________________________
Layer (type)                 Output Shape              Param #   
=================================================================
mobilenetv2_1.00_224 (Functi (None, 7, 7, 1280)        2257984   
_________________________________________________________________
global_average_pooling2d_2 ( (None, 1280)              0         
_________________________________________________________________
dropout_2 (Dropout)          (None, 1280)              0         
_________________________________________________________________
dense_1 (Dense)              (None, 3)                 3843      
=================================================================
Total params: 2,261,827
Trainable params: 3,843
Non-trainable params: 2,257,984

训练后我使用

转换模型
model = tf.keras.models.load_model('model.h5')
converter = tf.lite.TFLiteConverter.from_keras_model(model)
tflite_model = converter.convert()
open(f"myModel.tflite", "wb").write(tflite_model)

对于 android 代码如下:

        make_prediction.setOnClickListener(View.OnClickListener {
            var resized = Bitmap.createScaledBitmap(bitmap, 224, 224, true)
            val model = MyModel.newInstance(this)
            var tbuffer = TensorImage.fromBitmap(resized)
            var byteBuffer = tbuffer.buffer

// Creates inputs for reference.
            val inputFeature0 = TensorBuffer.createFixedSize(intArrayOf(1, 224, 224, 3), DataType.FLOAT32)
            inputFeature0.loadBuffer(byteBuffer)

// Runs model inference and gets result.
            val outputs = model.process(inputFeature0)
            val outputFeature0 = outputs.outputFeature0AsTensorBuffer

            var max = getMax(outputFeature0.floatArray)

            text_view.setText(labels[max])

// Releases model resources if no longer used.
            model.close()
        })

但是每当我尝试 运行 我的应用程序时它就会关闭并且我在 logcat.

中收到此错误
java.lang.IllegalArgumentException: The size of byte buffer and the shape do not match.

如果我将图像的输入形状从 224 更改为 300,并在 300 输入形状上训练我的模型并插入到 android,我会遇到错误。

java.lang.IllegalArgumentException: Cannot convert between a TensorFlowLite buffer with 1080000 bytes and a Java Buffer with 150528 bytes

非常感谢任何形式的帮助。

像这样使用它:

make_prediction.setOnClickListener(View.OnClickListener {
            var resized = Bitmap.createScaledBitmap(bitmap, 224, 224, true)
            val model = MyModel.newInstance(this)
            var tImage = TensorImage(DataType.FLOAT32)
            var tensorImage = tImage.load(resized)
            var byteBuffer = tensorImage.buffer

// Creates inputs for reference.
            //val inputFeature0 = TensorBuffer.createFixedSize(intArrayOf(1, 224, 224, 3), DataType.FLOAT32)
            //inputFeature0.loadBuffer(byteBuffer)

// Runs model inference and gets result.
            val outputs = model.process(byteBuffer)
            val outputFeature0 = outputs.outputFeature0AsTensorBuffer

            var max = getMax(outputFeature0.floatArray)

            text_view.setText(labels[max])

// Releases model resources if no longer used.
            model.close()
        })

然后检查调试器是否问题仍然存在或者

val outputFeature0 = outputs.outputFeature0AsTensorBuffer 导致另一个。

如果需要更多帮助,请联系我