问题描述
import numpy as np
import pandas as pd
import tensorflow as tf
from sklearn.model_selection import train_test_split
from tensorflow.keras import layers
from tensorflow.keras.layers.experimental import preprocessing
import os
import matplotlib.pyplot as plt
import pandas as pd
import tensorflow as tf
from tensorflow import keras
from tensorflow.keras import layers
from tensorflow.keras.layers.experimental import preprocessing
import seaborn as sns
import numpy as np
pip install -q -U keras-tuner
import keras_tuner as kt
df = pd.read_csv("/content/credit-approval_csv.csv",delimiter=',')
train,test = train_test_split(df,test_size=0.2)
train,val = train_test_split(train,test_size=0.2)
print(len(train),'train examples')
print(len(val),'validation examples')
print(len(test),'test examples')
def df_to_dataset(dataframe,shuffle=True,batch_size=32):
dataframe = df.copy()
labels = dataframe.pop('class')
ds = tf.data.Dataset.from_tensor_slices((dict(dataframe),labels))
if shuffle:
ds = ds.shuffle(buffer_size=len(dataframe))
ds = ds.batch(batch_size)
ds = ds.prefetch(batch_size)
return ds
def get_normalization_layer(name,dataset):
# Create a normalization layer for our feature.
normalizer = preprocessing.normalization(axis=None)
# Prepare a Dataset that only yields our feature.
feature_ds = dataset.map(lambda x,y: x[name])
# Learn the statistics of the data.
normalizer.adapt(feature_ds)
return normalizer
def get_category_encoding_layer(name,dataset,dtype,max_tokens=None):
# Create a StringLookup layer which will turn strings into integer indices
if dtype == 'string':
index = preprocessing.StringLookup(max_tokens=max_tokens)
else:
index = preprocessing.IntegerLookup(max_tokens=max_tokens)
# Prepare a Dataset that only yields our feature
feature_ds = dataset.map(lambda x,y: x[name])
# Learn the set of possible values and assign them a fixed integer index.
index.adapt(feature_ds)
# Create a discretization for our integer indices.
encoder = preprocessing.CategoryEncoding(num_tokens=index.vocabulary_size())
# Apply one-hot encoding to our indices. The lambda function captures the
# layer so we can use them,or include them in the functional model later.
return lambda feature: encoder(index(feature))
batch_size = 256
train_ds = df_to_dataset(train,batch_size=batch_size)
val_ds = df_to_dataset(val,shuffle=False,batch_size=batch_size)
test_ds = df_to_dataset(test,batch_size=batch_size)
all_inputs = []
encoded_features = []
# Numeric features.
for header in ['A2','A3','A8','A11','A14','A15']:
numeric_col = tf.keras.Input(shape=(1,),name=header)
normalization_layer = get_normalization_layer(header,train_ds)
encoded_numeric_col = normalization_layer(numeric_col)
all_inputs.append(numeric_col)
encoded_features.append(encoded_numeric_col)
# Categorical features encoded as string.
categorical_cols = ['A13','A12','A10','A9','A7','A6','A5','A4','A1']
for header in categorical_cols:
categorical_col = tf.keras.Input(shape=(1,name=header,dtype='string')
encoding_layer = get_category_encoding_layer(header,train_ds,dtype='string',max_tokens=5)
encoded_categorical_col = encoding_layer(categorical_col)
all_inputs.append(categorical_col)
encoded_features.append(encoded_categorical_col)
def build_model(hp):
hp_units = hp.Int('units',min_value=1,max_value=1512,step=32)
all_features = tf.keras.layers.concatenate(encoded_features)
dense = layers.Dense(units=hp_units,activation="relu")
x = dense(all_features)
x = layers.Dense(units=hp_units,activation="relu")(x)
x = layers.Dense(units=hp_units,activation="relu")(x)
x = layers.Dropout(rate=0.5)(x)
outputs = layers.Dense(units=hp_units)(x)
model = tf.keras.Model(all_inputs,outputs)
hp_learning_rate = hp.Choice('learning_rate',values=[1e-2,1e-3,1e-4])
optimizer = hp.Choice("optimizer",["adam","sgd","RMSprop"])
loss = hp.Choice("loss",["BinaryCrossentropy","CategoricalCrossentropy","SparseCategoricalCrossentropy"])
model.compile(optimizer,loss,metrics=['accuracy'])
return model
tuner = kt.Hyperband(build_model,objective='val_accuracy',max_epochs=10,factor=3,hyperband_iterations=1,directory='my_dir',project_name='intro_to_kt',overwrite=True)
tuner.search(train_ds,epochs=50,validation_data=val_ds)
这是发生以下错误时:
learning_rate |0.001 |?
optimizer |RMSprop |?
loss |CategoricalCros...|?
tuner/epochs |2 |?
tuner/initial_e...|0 |?
tuner/bracket |2 |?
tuner/round |0 |?
Epoch 1/2
---------------------------------------------------------------------------
ValueError Traceback (most recent call last)
<ipython-input-49-848b8addbd7e> in <module>()
----> 1 tuner.search(train_ds,validation_data=val_ds)
13 frames
/usr/local/lib/python3.7/dist-packages/tensorflow/python/framework/func_graph.py in wrapper(*args,**kwargs)
984 except Exception as e: # pylint:disable=broad-except
985 if hasattr(e,"ag_error_Metadata"):
--> 986 raise e.ag_error_Metadata.to_exception(e)
987 else:
988 raise
ValueError: in user code:
/usr/local/lib/python3.7/dist-packages/tensorflow/python/keras/engine/training.py:855 train_function *
return step_function(self,iterator)
/usr/local/lib/python3.7/dist-packages/tensorflow/python/keras/engine/training.py:845 step_function **
outputs = model.distribute_strategy.run(run_step,args=(data,))
/usr/local/lib/python3.7/dist-packages/tensorflow/python/distribute/distribute_lib.py:1285 run
return self._extended.call_for_each_replica(fn,args=args,kwargs=kwargs)
/usr/local/lib/python3.7/dist-packages/tensorflow/python/distribute/distribute_lib.py:2833 call_for_each_replica
return self._call_for_each_replica(fn,args,kwargs)
/usr/local/lib/python3.7/dist-packages/tensorflow/python/distribute/distribute_lib.py:3608 _call_for_each_replica
return fn(*args,**kwargs)
/usr/local/lib/python3.7/dist-packages/tensorflow/python/keras/engine/training.py:838 run_step **
outputs = model.train_step(data)
/usr/local/lib/python3.7/dist-packages/tensorflow/python/keras/engine/training.py:797 train_step
y,y_pred,sample_weight,regularization_losses=self.losses)
/usr/local/lib/python3.7/dist-packages/tensorflow/python/keras/engine/compile_utils.py:204 __call__
loss_value = loss_obj(y_t,y_p,sample_weight=sw)
/usr/local/lib/python3.7/dist-packages/tensorflow/python/keras/losses.py:155 __call__
losses = call_fn(y_true,y_pred)
/usr/local/lib/python3.7/dist-packages/tensorflow/python/keras/losses.py:259 call **
return ag_fn(y_true,**self._fn_kwargs)
/usr/local/lib/python3.7/dist-packages/tensorflow/python/util/dispatch.py:206 wrapper
return target(*args,**kwargs)
/usr/local/lib/python3.7/dist-packages/tensorflow/python/keras/losses.py:1644 categorical_crossentropy
y_true,from_logits=from_logits)
/usr/local/lib/python3.7/dist-packages/tensorflow/python/util/dispatch.py:206 wrapper
return target(*args,**kwargs)
/usr/local/lib/python3.7/dist-packages/tensorflow/python/keras/backend.py:4862 categorical_crossentropy
target.shape.assert_is_compatible_with(output.shape)
/usr/local/lib/python3.7/dist-packages/tensorflow/python/framework/tensor_shape.py:1161 assert_is_compatible_with
raise ValueError("Shapes %s and %s are incompatible" % (self,other))
ValueError: Shapes (None,1) and (None,609) are incompatible
解决方法
暂无找到可以解决该程序问题的有效方法,小编努力寻找整理中!
如果你已经找到好的解决方法,欢迎将解决方案带上本链接一起发送给小编。
小编邮箱:dio#foxmail.com (将#修改为@)