text
stringlengths 0
4.99k
|
---|
combined_images = tf.concat([generated_images, real_images], axis=0) |
# Assemble labels discriminating real from fake images |
labels = tf.concat( |
[tf.ones((batch_size, 1)), tf.zeros((real_images.shape[0], 1))], axis=0 |
) |
# Add random noise to the labels - important trick! |
labels += 0.05 * tf.random.uniform(labels.shape) |
# Train the discriminator |
with tf.GradientTape() as tape: |
predictions = discriminator(combined_images) |
d_loss = loss_fn(labels, predictions) |
grads = tape.gradient(d_loss, discriminator.trainable_weights) |
d_optimizer.apply_gradients(zip(grads, discriminator.trainable_weights)) |
# Sample random points in the latent space |
random_latent_vectors = tf.random.normal(shape=(batch_size, latent_dim)) |
# Assemble labels that say "all real images" |
misleading_labels = tf.zeros((batch_size, 1)) |
# Train the generator (note that we should *not* update the weights |
# of the discriminator)! |
with tf.GradientTape() as tape: |
predictions = discriminator(generator(random_latent_vectors)) |
g_loss = loss_fn(misleading_labels, predictions) |
grads = tape.gradient(g_loss, generator.trainable_weights) |
g_optimizer.apply_gradients(zip(grads, generator.trainable_weights)) |
return d_loss, g_loss, generated_images |
Let's train our GAN, by repeatedly calling train_step on batches of images. |
Since our discriminator and generator are convnets, you're going to want to run this code on a GPU. |
import os |
# Prepare the dataset. We use both the training & test MNIST digits. |
batch_size = 64 |
(x_train, _), (x_test, _) = keras.datasets.mnist.load_data() |
all_digits = np.concatenate([x_train, x_test]) |
all_digits = all_digits.astype("float32") / 255.0 |
all_digits = np.reshape(all_digits, (-1, 28, 28, 1)) |
dataset = tf.data.Dataset.from_tensor_slices(all_digits) |
dataset = dataset.shuffle(buffer_size=1024).batch(batch_size) |
epochs = 1 # In practice you need at least 20 epochs to generate nice digits. |
save_dir = "./" |
for epoch in range(epochs): |
print("\nStart epoch", epoch) |
for step, real_images in enumerate(dataset): |
# Train the discriminator & generator on one batch of real images. |
d_loss, g_loss, generated_images = train_step(real_images) |
# Logging. |
if step % 200 == 0: |
# Print metrics |
print("discriminator loss at step %d: %.2f" % (step, d_loss)) |
print("adversarial loss at step %d: %.2f" % (step, g_loss)) |
# Save one generated image |
img = tf.keras.preprocessing.image.array_to_img( |
generated_images[0] * 255.0, scale=False |
) |
img.save(os.path.join(save_dir, "generated_img" + str(step) + ".png")) |
# To limit execution time we stop after 10 steps. |
# Remove the lines below to actually train the model! |
if step > 10: |
break |
Start epoch 0 |
discriminator loss at step 0: 0.70 |
adversarial loss at step 0: 0.68 |
That's it! You'll get nice-looking fake MNIST digits after just ~30s of training on the Colab GPU.Serialization and saving |
Authors: Kathy Wu, Francois Chollet |
Date created: 2020/04/28 |
Last modified: 2020/04/28 |
Description: Complete guide to saving & serializing models. |
View in Colab • GitHub source |
Introduction |
A Keras model consists of multiple components: |
The architecture, or configuration, which specifies what layers the model contain, and how they're connected. |
A set of weights values (the "state of the model"). |
An optimizer (defined by compiling the model). |
A set of losses and metrics (defined by compiling the model or calling add_loss() or add_metric()). |
The Keras API makes it possible to save all of these pieces to disk at once, or to only selectively save some of them: |
Saving everything into a single archive in the TensorFlow SavedModel format (or in the older Keras H5 format). This is the standard practice. |
Saving the architecture / configuration only, typically as a JSON file. |
Saving the weights values only. This is generally used when training the model. |
Let's take a look at each of these options. When would you use one or the other, and how do they work? |
How to save and load a model |
If you only have 10 seconds to read this guide, here's what you need to know. |
Saving a Keras model: |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.