text
stringlengths 0
4.99k
|
---|
combined_images = tf.concat([generated_images, real_images], axis=0)
|
# Assemble labels discriminating real from fake images
|
labels = tf.concat(
|
[tf.ones((batch_size, 1)), tf.zeros((real_images.shape[0], 1))], axis=0
|
)
|
# Add random noise to the labels - important trick!
|
labels += 0.05 * tf.random.uniform(labels.shape)
|
# Train the discriminator
|
with tf.GradientTape() as tape:
|
predictions = discriminator(combined_images)
|
d_loss = loss_fn(labels, predictions)
|
grads = tape.gradient(d_loss, discriminator.trainable_weights)
|
d_optimizer.apply_gradients(zip(grads, discriminator.trainable_weights))
|
# Sample random points in the latent space
|
random_latent_vectors = tf.random.normal(shape=(batch_size, latent_dim))
|
# Assemble labels that say "all real images"
|
misleading_labels = tf.zeros((batch_size, 1))
|
# Train the generator (note that we should *not* update the weights
|
# of the discriminator)!
|
with tf.GradientTape() as tape:
|
predictions = discriminator(generator(random_latent_vectors))
|
g_loss = loss_fn(misleading_labels, predictions)
|
grads = tape.gradient(g_loss, generator.trainable_weights)
|
g_optimizer.apply_gradients(zip(grads, generator.trainable_weights))
|
return d_loss, g_loss, generated_images
|
Let's train our GAN, by repeatedly calling train_step on batches of images.
|
Since our discriminator and generator are convnets, you're going to want to run this code on a GPU.
|
import os
|
# Prepare the dataset. We use both the training & test MNIST digits.
|
batch_size = 64
|
(x_train, _), (x_test, _) = keras.datasets.mnist.load_data()
|
all_digits = np.concatenate([x_train, x_test])
|
all_digits = all_digits.astype("float32") / 255.0
|
all_digits = np.reshape(all_digits, (-1, 28, 28, 1))
|
dataset = tf.data.Dataset.from_tensor_slices(all_digits)
|
dataset = dataset.shuffle(buffer_size=1024).batch(batch_size)
|
epochs = 1 # In practice you need at least 20 epochs to generate nice digits.
|
save_dir = "./"
|
for epoch in range(epochs):
|
print("\nStart epoch", epoch)
|
for step, real_images in enumerate(dataset):
|
# Train the discriminator & generator on one batch of real images.
|
d_loss, g_loss, generated_images = train_step(real_images)
|
# Logging.
|
if step % 200 == 0:
|
# Print metrics
|
print("discriminator loss at step %d: %.2f" % (step, d_loss))
|
print("adversarial loss at step %d: %.2f" % (step, g_loss))
|
# Save one generated image
|
img = tf.keras.preprocessing.image.array_to_img(
|
generated_images[0] * 255.0, scale=False
|
)
|
img.save(os.path.join(save_dir, "generated_img" + str(step) + ".png"))
|
# To limit execution time we stop after 10 steps.
|
# Remove the lines below to actually train the model!
|
if step > 10:
|
break
|
Start epoch 0
|
discriminator loss at step 0: 0.70
|
adversarial loss at step 0: 0.68
|
That's it! You'll get nice-looking fake MNIST digits after just ~30s of training on the Colab GPU.Serialization and saving
|
Authors: Kathy Wu, Francois Chollet
|
Date created: 2020/04/28
|
Last modified: 2020/04/28
|
Description: Complete guide to saving & serializing models.
|
View in Colab • GitHub source
|
Introduction
|
A Keras model consists of multiple components:
|
The architecture, or configuration, which specifies what layers the model contain, and how they're connected.
|
A set of weights values (the "state of the model").
|
An optimizer (defined by compiling the model).
|
A set of losses and metrics (defined by compiling the model or calling add_loss() or add_metric()).
|
The Keras API makes it possible to save all of these pieces to disk at once, or to only selectively save some of them:
|
Saving everything into a single archive in the TensorFlow SavedModel format (or in the older Keras H5 format). This is the standard practice.
|
Saving the architecture / configuration only, typically as a JSON file.
|
Saving the weights values only. This is generally used when training the model.
|
Let's take a look at each of these options. When would you use one or the other, and how do they work?
|
How to save and load a model
|
If you only have 10 seconds to read this guide, here's what you need to know.
|
Saving a Keras model:
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.