import tensorflow as tf
import numpy as np
from sklearn.model_selection import train_test_split
np.random.seed(4213)
data = np.random.randint(low=1,high=29, size=(500, 160, 160, 10))
labels = np.random.randint(low=0,high=5, size=(500, 160, 160))
nclass = len(np.unique(labels))
print (nclass)
samples, width, height, nbands = data.shape
X_train, X_test, y_train, y_test = train_test_split(data, labels, test_size=0.25, random_state=421)
print (X_train.shape)
print (y_train.shape)
arch = tf.keras.applications.VGG16(input_shape=[width, height, nbands],
include_top=False,
weights=None)
model = tf.keras.Sequential()
model.add(arch)
model.add(tf.keras.layers.Flatten())
model.add(tf.keras.layers.Dense(nclass))
model.compile(optimizer = tf.keras.optimizers.Adam(0.0001),
loss=tf.keras.losses.SparseCategoricalCrossentropy(),
metrics=[tf.keras.metrics.SparseCategoricalAccuracy()])
model.fit(X_train,
y_train,
epochs=3,
batch_size=32,
verbose=2)
res = model.predict(X_test)
print(res.shape)
为semantic segmentation
运行上述代码时,发生了I get异常:
InvalidArgumentError
Incompatible shapes: [32,160,160] vs. [32]
[[node Equal (defined at c...:38) ]] [Op:__inference_train_function_1815]
tensorflow.python.framework.errors_impl.InvalidArgumentError
您的问题来自最后一层的大小(为了避免这些错误,最好对
N_IMAGES
、WIDTH
、HEIGHT
、N_CHANNELS
和N_CLASSES
使用python常量):对于图像分类
您应该为每个图像指定一个标签。尝试切换
labels
:对于语义分段
确保分类器(网络的最后一层)的大小相应。在这种情况下,每像素需要1个类:
这是你能得到的最简单的。相反,您可以设置多个反褶积层作为分类器,或者甚至可以翻转
arch
体系结构并使用它生成分类结果。正交地,您可以对标签执行one_hot
编码,从而将标签扩展为N_CLASSES
因子,有效地增加最后一层中的神经元数量相关问题 更多 >
编程相关推荐