text
stringlengths 0
4.99k
|
---|
) |
self.kernel_2_3 = self.add_weight( |
shape=(i2, i3, self.unit_2, self.unit_3), |
initializer="uniform", |
name="kernel_2_3", |
) |
def call(self, inputs, states): |
# inputs should be in [(batch, input_1), (batch, input_2, input_3)] |
# state should be in shape [(batch, unit_1), (batch, unit_2, unit_3)] |
input_1, input_2 = tf.nest.flatten(inputs) |
s1, s2 = states |
output_1 = tf.matmul(input_1, self.kernel_1) |
output_2_3 = tf.einsum("bij,ijkl->bkl", input_2, self.kernel_2_3) |
state_1 = s1 + output_1 |
state_2_3 = s2 + output_2_3 |
output = (output_1, output_2_3) |
new_states = (state_1, state_2_3) |
return output, new_states |
def get_config(self): |
return {"unit_1": self.unit_1, "unit_2": unit_2, "unit_3": self.unit_3} |
Build a RNN model with nested input/output |
Let's build a Keras model that uses a keras.layers.RNN layer and the custom cell we just defined. |
unit_1 = 10 |
unit_2 = 20 |
unit_3 = 30 |
i1 = 32 |
i2 = 64 |
i3 = 32 |
batch_size = 64 |
num_batches = 10 |
timestep = 50 |
cell = NestedCell(unit_1, unit_2, unit_3) |
rnn = keras.layers.RNN(cell) |
input_1 = keras.Input((None, i1)) |
input_2 = keras.Input((None, i2, i3)) |
outputs = rnn((input_1, input_2)) |
model = keras.models.Model([input_1, input_2], outputs) |
model.compile(optimizer="adam", loss="mse", metrics=["accuracy"]) |
Train the model with randomly generated data |
Since there isn't a good candidate dataset for this model, we use random Numpy data for demonstration. |
input_1_data = np.random.random((batch_size * num_batches, timestep, i1)) |
input_2_data = np.random.random((batch_size * num_batches, timestep, i2, i3)) |
target_1_data = np.random.random((batch_size * num_batches, unit_1)) |
target_2_data = np.random.random((batch_size * num_batches, unit_2, unit_3)) |
input_data = [input_1_data, input_2_data] |
target_data = [target_1_data, target_2_data] |
model.fit(input_data, target_data, batch_size=batch_size) |
10/10 [==============================] - 4s 263ms/step - loss: 0.9004 - rnn_1_loss: 0.3103 - rnn_1_1_loss: 0.5902 - rnn_1_accuracy: 0.1403 - rnn_1_1_accuracy: 0.0335 |
<tensorflow.python.keras.callbacks.History at 0x154ce1990> |
With the Keras keras.layers.RNN layer, You are only expected to define the math logic for individual step within the sequence, and the keras.layers.RNN layer will handle the sequence iteration for you. It's an incredibly powerful way to quickly prototype new kinds of RNNs (e.g. a LSTM variant). |
For more details, please visit the API docs.Writing a training loop from scratch |
Author: fchollet |
Date created: 2019/03/01 |
Last modified: 2020/04/15 |
Description: Complete guide to writing low-level training & evaluation loops. |
View in Colab • GitHub source |
Setup |
import tensorflow as tf |
from tensorflow import keras |
from tensorflow.keras import layers |
import numpy as np |
Introduction |
Keras provides default training and evaluation loops, fit() and evaluate(). Their usage is covered in the guide Training & evaluation with the built-in methods. |
If you want to customize the learning algorithm of your model while still leveraging the convenience of fit() (for instance, to train a GAN using fit()), you can subclass the Model class and implement your own train_step() method, which is called repeatedly during fit(). This is covered in the guide Customizing what happens in fit(). |
Now, if you want very low-level control over training & evaluation, you should write your own training & evaluation loops from scratch. This is what this guide is about. |
Using the GradientTape: a first end-to-end example |
Calling a model inside a GradientTape scope enables you to retrieve the gradients of the trainable weights of the layer with respect to a loss value. Using an optimizer instance, you can use these gradients to update these variables (which you can retrieve using model.trainable_weights). |
Let's consider a simple MNIST model: |
inputs = keras.Input(shape=(784,), name="digits") |
x1 = layers.Dense(64, activation="relu")(inputs) |
x2 = layers.Dense(64, activation="relu")(x1) |
outputs = layers.Dense(10, name="predictions")(x2) |
model = keras.Model(inputs=inputs, outputs=outputs) |
Let's train it using mini-batch gradient with a custom training loop. |
First, we're going to need an optimizer, a loss function, and a dataset: |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.