text
stringlengths 0
4.99k
|
---|
0: 1.0,
|
1: 1.0,
|
2: 1.0,
|
3: 1.0,
|
4: 1.0,
|
# Set weight "2" for class "5",
|
# making this class 2x more important
|
5: 2.0,
|
6: 1.0,
|
7: 1.0,
|
8: 1.0,
|
9: 1.0,
|
}
|
print("Fit with class weight")
|
model = get_compiled_model()
|
model.fit(x_train, y_train, class_weight=class_weight, batch_size=64, epochs=1)
|
Fit with class weight
|
782/782 [==============================] - 1s 933us/step - loss: 0.6334 - sparse_categorical_accuracy: 0.8297
|
<tensorflow.python.keras.callbacks.History at 0x14e20f990>
|
Sample weights
|
For fine grained control, or if you are not building a classifier, you can use "sample weights".
|
When training from NumPy data: Pass the sample_weight argument to Model.fit().
|
When training from tf.data or any other sort of iterator: Yield (input_batch, label_batch, sample_weight_batch) tuples.
|
A "sample weights" array is an array of numbers that specify how much weight each sample in a batch should have in computing the total loss. It is commonly used in imbalanced classification problems (the idea being to give more weight to rarely-seen classes).
|
When the weights used are ones and zeros, the array can be used as a mask for the loss function (entirely discarding the contribution of certain samples to the total loss).
|
sample_weight = np.ones(shape=(len(y_train),))
|
sample_weight[y_train == 5] = 2.0
|
print("Fit with sample weight")
|
model = get_compiled_model()
|
model.fit(x_train, y_train, sample_weight=sample_weight, batch_size=64, epochs=1)
|
Fit with sample weight
|
782/782 [==============================] - 1s 899us/step - loss: 0.6337 - sparse_categorical_accuracy: 0.8355
|
<tensorflow.python.keras.callbacks.History at 0x14e3538d0>
|
Here's a matching Dataset example:
|
sample_weight = np.ones(shape=(len(y_train),))
|
sample_weight[y_train == 5] = 2.0
|
# Create a Dataset that includes sample weights
|
# (3rd element in the return tuple).
|
train_dataset = tf.data.Dataset.from_tensor_slices((x_train, y_train, sample_weight))
|
# Shuffle and slice the dataset.
|
train_dataset = train_dataset.shuffle(buffer_size=1024).batch(64)
|
model = get_compiled_model()
|
model.fit(train_dataset, epochs=1)
|
782/782 [==============================] - 1s 1ms/step - loss: 0.6539 - sparse_categorical_accuracy: 0.8364
|
<tensorflow.python.keras.callbacks.History at 0x14e49c390>
|
Passing data to multi-input, multi-output models
|
In the previous examples, we were considering a model with a single input (a tensor of shape (764,)) and a single output (a prediction tensor of shape (10,)). But what about models that have multiple inputs or outputs?
|
Consider the following model, which has an image input of shape (32, 32, 3) (that's (height, width, channels)) and a time series input of shape (None, 10) (that's (timesteps, features)). Our model will have two outputs computed from the combination of these inputs: a "score" (of shape (1,)) and a probability distribution over five classes (of shape (5,)).
|
image_input = keras.Input(shape=(32, 32, 3), name="img_input")
|
timeseries_input = keras.Input(shape=(None, 10), name="ts_input")
|
x1 = layers.Conv2D(3, 3)(image_input)
|
x1 = layers.GlobalMaxPooling2D()(x1)
|
x2 = layers.Conv1D(3, 3)(timeseries_input)
|
x2 = layers.GlobalMaxPooling1D()(x2)
|
x = layers.concatenate([x1, x2])
|
score_output = layers.Dense(1, name="score_output")(x)
|
class_output = layers.Dense(5, name="class_output")(x)
|
model = keras.Model(
|
inputs=[image_input, timeseries_input], outputs=[score_output, class_output]
|
)
|
Let's plot this model, so you can clearly see what we're doing here (note that the shapes shown in the plot are batch shapes, rather than per-sample shapes).
|
keras.utils.plot_model(model, "multi_input_and_output_model.png", show_shapes=True)
|
png
|
At compilation time, we can specify different losses to different outputs, by passing the loss functions as a list:
|
model.compile(
|
optimizer=keras.optimizers.RMSprop(1e-3),
|
loss=[keras.losses.MeanSquaredError(), keras.losses.CategoricalCrossentropy()],
|
)
|
If we only passed a single loss function to the model, the same loss function would be applied to every output (which is not appropriate here).
|
Likewise for metrics:
|
model.compile(
|
optimizer=keras.optimizers.RMSprop(1e-3),
|
loss=[keras.losses.MeanSquaredError(), keras.losses.CategoricalCrossentropy()],
|
metrics=[
|
[
|
keras.metrics.MeanAbsolutePercentageError(),
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.