The Keras functional API is a way to create models that are more flexible than the sequential API. The functional API can handle models with non-linear topology, shared layers, and even multiple inputs or outputs.
The main idea is that a deep learning model is usually a directed acyclic graph (DAG) of layers. So the functional API is a way to build graphs of layers.
Consider the following model:
(input: 784-dimensional vectors)
↧
[Dense (64 units, relu activation)]
↧
[Dense (64 units, relu activation)]
↧
[Dense (10 units, softmax activation)]
↧
(output: logits of a probability distribution over 10 classes)
This is a basic graph with three layers. To build this model using the functional API, start by creating an input node:
The shape of the data is set as a 784-dimensional vector. The batch size is always omitted since only the shape of each sample is specified.
If, for example, you have an image input with a shape of
(32, 32, 3)
, you would use:
The inputs
that is returned contains information about
the shape and dtype
of the input data that you feed to your
model. Here’s the shape:
## shape(NA, 784)
Here’s the dtype:
## [1] "float32"
You create a new node in the graph of layers by calling a layer on
this inputs
object:
The “layer call” action is like drawing an arrow from “inputs” to
this layer you created. You’re “passing” the inputs to the
dense
layer, and you get x
as the output.
Let’s add a few more layers to the graph of layers:
At this point, you can create a Model
by specifying its
inputs and outputs in the graph of layers:
Let’s check out what the model summary looks like:
## [1mModel: "mnist_model"[0m
## ┏━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━┓
## ┃[1m [0m[1mLayer (type) [0m[1m [0m┃[1m [0m[1mOutput Shape [0m[1m [0m┃[1m [0m[1m Param #[0m[1m [0m┃
## ┡━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━┩
## │ input_layer ([38;5;33mInputLayer[0m) │ ([38;5;45mNone[0m, [38;5;34m784[0m) │ [38;5;34m0[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ dense ([38;5;33mDense[0m) │ ([38;5;45mNone[0m, [38;5;34m64[0m) │ [38;5;34m50,240[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ dense_1 ([38;5;33mDense[0m) │ ([38;5;45mNone[0m, [38;5;34m64[0m) │ [38;5;34m4,160[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ dense_2 ([38;5;33mDense[0m) │ ([38;5;45mNone[0m, [38;5;34m10[0m) │ [38;5;34m650[0m │
## └─────────────────────────────────┴────────────────────────┴───────────────┘
## [1m Total params: [0m[38;5;34m55,050[0m (215.04 KB)
## [1m Trainable params: [0m[38;5;34m55,050[0m (215.04 KB)
## [1m Non-trainable params: [0m[38;5;34m0[0m (0.00 B)
You can also plot the model as a graph:
And, optionally, display the input and output shapes of each layer in the plotted graph:
This figure and the code are almost identical. In the code version, the connection arrows are replaced by the call operation.
A “graph of layers” is an intuitive mental image for a deep learning model, and the functional API is a way to create models that closely mirrors this.
Training, evaluation, and inference work exactly in the same way for
models built using the functional API as for Sequential
models.
The Model
class offers a built-in training loop (the
fit()
method) and a built-in evaluation loop (the
evaluate()
method). Note that you can easily customize these loops to
implement training routines beyond supervised learning (e.g. GANs).
Here, load the MNIST image data, reshape it into vectors, fit the model on the data (while monitoring performance on a validation split), then evaluate the model on the test data:
c(c(x_train, y_train), c(x_test, y_test)) %<-% dataset_mnist()
x_train <- array_reshape(x_train, c(60000, 784)) / 255
x_test <- array_reshape(x_test, c(10000, 784)) / 255
model |> compile(
loss = loss_sparse_categorical_crossentropy(from_logits = TRUE),
optimizer = optimizer_rmsprop(),
metrics = "accuracy"
)
history <- model |> fit(
x_train, y_train, batch_size = 64, epochs = 2, validation_split = 0.2
)
## Epoch 1/2
## 750/750 - 2s - 3ms/step - accuracy: 0.8979 - loss: 0.3540 - val_accuracy: 0.9448 - val_loss: 0.1903
## Epoch 2/2
## 750/750 - 1s - 784us/step - accuracy: 0.9509 - loss: 0.1635 - val_accuracy: 0.9597 - val_loss: 0.1397
## 313/313 - 0s - 1ms/step - accuracy: 0.9595 - loss: 0.1328
## Test loss: 0.132778
## Test accuracy: 0.9595
For further reading, see the training and evaluation guide.
Saving the model and serialization work the same way for models built
using the functional API as they do for Sequential
models.
The standard way to save a functional model is to call
model.save()
to save the entire model as a single file. You
can later recreate the same model from this file, even if the code that
built the model is no longer available.
This saved file includes the: - model architecture - model weight
values (that were learned during training) - model training config, if
any (as passed to compile()
) - optimizer and its state, if
any (to restart training where you left off)
model |> save_model("my_model.keras")
rm(model)
# Recreate the exact same model purely from the file:
model <- load_model("my_model.keras")
For details, read the model serialization & saving guide.
In the functional API, models are created by specifying their inputs and outputs in a graph of layers. That means that a single graph of layers can be used to generate multiple models.
In the example below, you use the same stack of layers to instantiate
two models: an encoder
model that turns image inputs into
16-dimensional vectors, and an end-to-end autoencoder
model
for training.
encoder_input <- keras_input(shape = c(28, 28, 1), name="img")
encoder_output <- encoder_input |>
layer_conv_2d(16, 3, activation = "relu") |>
layer_conv_2d(32, 3, activation = "relu") |>
layer_max_pooling_2d(3) |>
layer_conv_2d(32, 3, activation = "relu") |>
layer_conv_2d(16, 3, activation = "relu") |>
layer_global_max_pooling_2d()
encoder <- keras_model(encoder_input, encoder_output, name="encoder")
summary(encoder)
## [1mModel: "encoder"[0m
## ┏━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━┓
## ┃[1m [0m[1mLayer (type) [0m[1m [0m┃[1m [0m[1mOutput Shape [0m[1m [0m┃[1m [0m[1m Param #[0m[1m [0m┃
## ┡━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━┩
## │ img ([38;5;33mInputLayer[0m) │ ([38;5;45mNone[0m, [38;5;34m28[0m, [38;5;34m28[0m, [38;5;34m1[0m) │ [38;5;34m0[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ conv2d ([38;5;33mConv2D[0m) │ ([38;5;45mNone[0m, [38;5;34m26[0m, [38;5;34m26[0m, [38;5;34m16[0m) │ [38;5;34m160[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ conv2d_1 ([38;5;33mConv2D[0m) │ ([38;5;45mNone[0m, [38;5;34m24[0m, [38;5;34m24[0m, [38;5;34m32[0m) │ [38;5;34m4,640[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ max_pooling2d ([38;5;33mMaxPooling2D[0m) │ ([38;5;45mNone[0m, [38;5;34m8[0m, [38;5;34m8[0m, [38;5;34m32[0m) │ [38;5;34m0[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ conv2d_2 ([38;5;33mConv2D[0m) │ ([38;5;45mNone[0m, [38;5;34m6[0m, [38;5;34m6[0m, [38;5;34m32[0m) │ [38;5;34m9,248[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ conv2d_3 ([38;5;33mConv2D[0m) │ ([38;5;45mNone[0m, [38;5;34m4[0m, [38;5;34m4[0m, [38;5;34m16[0m) │ [38;5;34m4,624[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ global_max_pooling2d │ ([38;5;45mNone[0m, [38;5;34m16[0m) │ [38;5;34m0[0m │
## │ ([38;5;33mGlobalMaxPooling2D[0m) │ │ │
## └─────────────────────────────────┴────────────────────────┴───────────────┘
## [1m Total params: [0m[38;5;34m18,672[0m (72.94 KB)
## [1m Trainable params: [0m[38;5;34m18,672[0m (72.94 KB)
## [1m Non-trainable params: [0m[38;5;34m0[0m (0.00 B)
decoder_output <- encoder_output |>
layer_reshape(c(4, 4, 1)) |>
layer_conv_2d_transpose(16, 3, activation = "relu") |>
layer_conv_2d_transpose(32, 3, activation = "relu") |>
layer_upsampling_2d(3) |>
layer_conv_2d_transpose(16, 3, activation = "relu") |>
layer_conv_2d_transpose(1, 3, activation = "relu")
autoencoder <- keras_model(encoder_input, decoder_output, name="autoencoder")
summary(autoencoder)
## [1mModel: "autoencoder"[0m
## ┏━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━┓
## ┃[1m [0m[1mLayer (type) [0m[1m [0m┃[1m [0m[1mOutput Shape [0m[1m [0m┃[1m [0m[1m Param #[0m[1m [0m┃
## ┡━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━┩
## │ img ([38;5;33mInputLayer[0m) │ ([38;5;45mNone[0m, [38;5;34m28[0m, [38;5;34m28[0m, [38;5;34m1[0m) │ [38;5;34m0[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ conv2d ([38;5;33mConv2D[0m) │ ([38;5;45mNone[0m, [38;5;34m26[0m, [38;5;34m26[0m, [38;5;34m16[0m) │ [38;5;34m160[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ conv2d_1 ([38;5;33mConv2D[0m) │ ([38;5;45mNone[0m, [38;5;34m24[0m, [38;5;34m24[0m, [38;5;34m32[0m) │ [38;5;34m4,640[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ max_pooling2d ([38;5;33mMaxPooling2D[0m) │ ([38;5;45mNone[0m, [38;5;34m8[0m, [38;5;34m8[0m, [38;5;34m32[0m) │ [38;5;34m0[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ conv2d_2 ([38;5;33mConv2D[0m) │ ([38;5;45mNone[0m, [38;5;34m6[0m, [38;5;34m6[0m, [38;5;34m32[0m) │ [38;5;34m9,248[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ conv2d_3 ([38;5;33mConv2D[0m) │ ([38;5;45mNone[0m, [38;5;34m4[0m, [38;5;34m4[0m, [38;5;34m16[0m) │ [38;5;34m4,624[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ global_max_pooling2d │ ([38;5;45mNone[0m, [38;5;34m16[0m) │ [38;5;34m0[0m │
## │ ([38;5;33mGlobalMaxPooling2D[0m) │ │ │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ reshape ([38;5;33mReshape[0m) │ ([38;5;45mNone[0m, [38;5;34m4[0m, [38;5;34m4[0m, [38;5;34m1[0m) │ [38;5;34m0[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ conv2d_transpose │ ([38;5;45mNone[0m, [38;5;34m6[0m, [38;5;34m6[0m, [38;5;34m16[0m) │ [38;5;34m160[0m │
## │ ([38;5;33mConv2DTranspose[0m) │ │ │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ conv2d_transpose_1 │ ([38;5;45mNone[0m, [38;5;34m8[0m, [38;5;34m8[0m, [38;5;34m32[0m) │ [38;5;34m4,640[0m │
## │ ([38;5;33mConv2DTranspose[0m) │ │ │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ up_sampling2d ([38;5;33mUpSampling2D[0m) │ ([38;5;45mNone[0m, [38;5;34m24[0m, [38;5;34m24[0m, [38;5;34m32[0m) │ [38;5;34m0[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ conv2d_transpose_2 │ ([38;5;45mNone[0m, [38;5;34m26[0m, [38;5;34m26[0m, [38;5;34m16[0m) │ [38;5;34m4,624[0m │
## │ ([38;5;33mConv2DTranspose[0m) │ │ │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ conv2d_transpose_3 │ ([38;5;45mNone[0m, [38;5;34m28[0m, [38;5;34m28[0m, [38;5;34m1[0m) │ [38;5;34m145[0m │
## │ ([38;5;33mConv2DTranspose[0m) │ │ │
## └─────────────────────────────────┴────────────────────────┴───────────────┘
## [1m Total params: [0m[38;5;34m28,241[0m (110.32 KB)
## [1m Trainable params: [0m[38;5;34m28,241[0m (110.32 KB)
## [1m Non-trainable params: [0m[38;5;34m0[0m (0.00 B)
Here, the decoding architecture is strictly symmetrical to the
encoding architecture, so the output shape is the same as the input
shape (28, 28, 1)
.
The reverse of a conv_2d
layer is a
conv_2d_transpose
layer, and the reverse of a
max_pooling_2d
layer is an upsampling_2d
layer.
You can treat any model as if it were a layer by invoking it on an
Input
or on the output of another layer. By calling a model
you aren’t just reusing the architecture of the model, you’re also
reusing its weights.
To see this in action, here’s a different take on the autoencoder example that creates an encoder model, a decoder model, and chains them in two calls to obtain the autoencoder model:
encoder_input <- keras_input(shape = c(28, 28, 1), name="img")
encoder_output <- encoder_input |>
layer_conv_2d(16, 3, activation = "relu") |>
layer_conv_2d(32, 3, activation = "relu") |>
layer_max_pooling_2d(3) |>
layer_conv_2d(32, 3, activation = "relu") |>
layer_conv_2d(16, 3, activation = "relu") |>
layer_global_max_pooling_2d()
encoder <- keras_model(encoder_input, encoder_output, name="encoder")
summary(encoder)
## [1mModel: "encoder"[0m
## ┏━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━┓
## ┃[1m [0m[1mLayer (type) [0m[1m [0m┃[1m [0m[1mOutput Shape [0m[1m [0m┃[1m [0m[1m Param #[0m[1m [0m┃
## ┡━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━┩
## │ img ([38;5;33mInputLayer[0m) │ ([38;5;45mNone[0m, [38;5;34m28[0m, [38;5;34m28[0m, [38;5;34m1[0m) │ [38;5;34m0[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ conv2d_4 ([38;5;33mConv2D[0m) │ ([38;5;45mNone[0m, [38;5;34m26[0m, [38;5;34m26[0m, [38;5;34m16[0m) │ [38;5;34m160[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ conv2d_5 ([38;5;33mConv2D[0m) │ ([38;5;45mNone[0m, [38;5;34m24[0m, [38;5;34m24[0m, [38;5;34m32[0m) │ [38;5;34m4,640[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ max_pooling2d_1 ([38;5;33mMaxPooling2D[0m) │ ([38;5;45mNone[0m, [38;5;34m8[0m, [38;5;34m8[0m, [38;5;34m32[0m) │ [38;5;34m0[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ conv2d_6 ([38;5;33mConv2D[0m) │ ([38;5;45mNone[0m, [38;5;34m6[0m, [38;5;34m6[0m, [38;5;34m32[0m) │ [38;5;34m9,248[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ conv2d_7 ([38;5;33mConv2D[0m) │ ([38;5;45mNone[0m, [38;5;34m4[0m, [38;5;34m4[0m, [38;5;34m16[0m) │ [38;5;34m4,624[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ global_max_pooling2d_1 │ ([38;5;45mNone[0m, [38;5;34m16[0m) │ [38;5;34m0[0m │
## │ ([38;5;33mGlobalMaxPooling2D[0m) │ │ │
## └─────────────────────────────────┴────────────────────────┴───────────────┘
## [1m Total params: [0m[38;5;34m18,672[0m (72.94 KB)
## [1m Trainable params: [0m[38;5;34m18,672[0m (72.94 KB)
## [1m Non-trainable params: [0m[38;5;34m0[0m (0.00 B)
decoder_input <- keras_input(shape = c(16), name = "encoded_img")
decoder_output <- decoder_input |>
layer_reshape(c(4, 4, 1)) |>
layer_conv_2d_transpose(16, 3, activation = "relu") |>
layer_conv_2d_transpose(32, 3, activation = "relu") |>
layer_upsampling_2d(3) |>
layer_conv_2d_transpose(16, 3, activation = "relu") |>
layer_conv_2d_transpose(1, 3, activation = "relu")
decoder <- keras_model(decoder_input, decoder_output, name = "decoder")
summary(decoder)
## [1mModel: "decoder"[0m
## ┏━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━┓
## ┃[1m [0m[1mLayer (type) [0m[1m [0m┃[1m [0m[1mOutput Shape [0m[1m [0m┃[1m [0m[1m Param #[0m[1m [0m┃
## ┡━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━┩
## │ encoded_img ([38;5;33mInputLayer[0m) │ ([38;5;45mNone[0m, [38;5;34m16[0m) │ [38;5;34m0[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ reshape_1 ([38;5;33mReshape[0m) │ ([38;5;45mNone[0m, [38;5;34m4[0m, [38;5;34m4[0m, [38;5;34m1[0m) │ [38;5;34m0[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ conv2d_transpose_4 │ ([38;5;45mNone[0m, [38;5;34m6[0m, [38;5;34m6[0m, [38;5;34m16[0m) │ [38;5;34m160[0m │
## │ ([38;5;33mConv2DTranspose[0m) │ │ │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ conv2d_transpose_5 │ ([38;5;45mNone[0m, [38;5;34m8[0m, [38;5;34m8[0m, [38;5;34m32[0m) │ [38;5;34m4,640[0m │
## │ ([38;5;33mConv2DTranspose[0m) │ │ │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ up_sampling2d_1 ([38;5;33mUpSampling2D[0m) │ ([38;5;45mNone[0m, [38;5;34m24[0m, [38;5;34m24[0m, [38;5;34m32[0m) │ [38;5;34m0[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ conv2d_transpose_6 │ ([38;5;45mNone[0m, [38;5;34m26[0m, [38;5;34m26[0m, [38;5;34m16[0m) │ [38;5;34m4,624[0m │
## │ ([38;5;33mConv2DTranspose[0m) │ │ │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ conv2d_transpose_7 │ ([38;5;45mNone[0m, [38;5;34m28[0m, [38;5;34m28[0m, [38;5;34m1[0m) │ [38;5;34m145[0m │
## │ ([38;5;33mConv2DTranspose[0m) │ │ │
## └─────────────────────────────────┴────────────────────────┴───────────────┘
## [1m Total params: [0m[38;5;34m9,569[0m (37.38 KB)
## [1m Trainable params: [0m[38;5;34m9,569[0m (37.38 KB)
## [1m Non-trainable params: [0m[38;5;34m0[0m (0.00 B)
autoencoder_input <- keras_input(shape = c(28, 28, 1), name = "img")
encoded_img <- encoder(autoencoder_input)
decoded_img <- decoder(encoded_img)
autoencoder <- keras_model(autoencoder_input, decoded_img,
name = "autoencoder")
summary(autoencoder)
## [1mModel: "autoencoder"[0m
## ┏━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━┓
## ┃[1m [0m[1mLayer (type) [0m[1m [0m┃[1m [0m[1mOutput Shape [0m[1m [0m┃[1m [0m[1m Param #[0m[1m [0m┃
## ┡━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━┩
## │ img ([38;5;33mInputLayer[0m) │ ([38;5;45mNone[0m, [38;5;34m28[0m, [38;5;34m28[0m, [38;5;34m1[0m) │ [38;5;34m0[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ encoder ([38;5;33mFunctional[0m) │ ([38;5;45mNone[0m, [38;5;34m16[0m) │ [38;5;34m18,672[0m │
## ├─────────────────────────────────┼────────────────────────┼───────────────┤
## │ decoder ([38;5;33mFunctional[0m) │ ([38;5;45mNone[0m, [38;5;34m28[0m, [38;5;34m28[0m, [38;5;34m1[0m) │ [38;5;34m9,569[0m │
## └─────────────────────────────────┴────────────────────────┴───────────────┘
## [1m Total params: [0m[38;5;34m28,241[0m (110.32 KB)
## [1m Trainable params: [0m[38;5;34m28,241[0m (110.32 KB)
## [1m Non-trainable params: [0m[38;5;34m0[0m (0.00 B)
As you can see, the model can be nested: a model can contain sub-models (since a model is just like a layer). A common use case for model nesting is ensembling. For example, here’s how to ensemble a set of models into a single model that averages their predictions:
get_model <- function() {
inputs <- keras_input(shape = 128)
outputs <- inputs |> layer_dense(1)
keras_model(inputs, outputs)
}
model1 <- get_model()
model2 <- get_model()
model3 <- get_model()
inputs <- keras_input(shape = 128)
y1 <- model1(inputs)
y2 <- model2(inputs)
y3 <- model3(inputs)
outputs <- layer_average(list(y1, y2, y3))
ensemble_model <- keras_model(inputs = inputs, outputs = outputs)
The functional API makes it easy to manipulate multiple inputs and
outputs. This cannot be handled with the Sequential
API.
For example, if you’re building a system for ranking customer issue tickets by priority and routing them to the correct department, then the model will have three inputs:
This model will have two outputs:
You can build this model in a few lines with the functional API:
num_tags <- 12 # Number of unique issue tags
num_words <- 10000 # Size of vocabulary obtained when preprocessing text data
num_departments <- 4 # Number of departments for predictions
title_input <- # Variable-length sequence of ints
keras_input(shape(NA), name = "title")
body_input <- # Variable-length sequence of ints
keras_input(shape(NA), name = "body")
tags_input <- # Binary vectors of size `num_tags`
keras_input(shape = num_tags, name = "tags")
# Embed each word in the title into a 64-dimensional vector
title_features <- layer_embedding(title_input, num_words, 64)
# Embed each word in the text into a 64-dimensional vector
body_features <- layer_embedding(body_input, num_words, 64)
# Reduce sequence of embedded words in the title
# into a single 128-dimensional vector
title_features <- layer_lstm(title_features, 128)
# Reduce sequence of embedded words in the body
# into a single 32-dimensional vector
body_features <- layer_lstm(body_features, 32)
# Merge all available features into a single large vector via concatenation
x <- layer_concatenate(title_features, body_features, tags_input)
# Stick a logistic regression for priority prediction on top of the features
priority_pred <- layer_dense(x, 1, name = "priority")
# Stick a department classifier on top of the features
department_pred <- layer_dense(x, num_departments, name = "department")
# Instantiate an end-to-end model predicting both priority and department
model <- keras_model(
inputs = list(title_input, body_input, tags_input),
outputs = list(priority = priority_pred, department = department_pred)
)
Now plot the model:
When compiling this model, you can assign different losses to each output. You can even assign different weights to each loss – to modulate their contribution to the total training loss.
model |> compile(
optimizer = optimizer_rmsprop(1e-3),
loss = list(
loss_binary_crossentropy(from_logits = TRUE),
loss_categorical_crossentropy(from_logits = TRUE)
),
loss_weights = c(1.0, 0.2)
)
Since the output layers have different names, you could also specify the losses and loss weights with the corresponding layer names:
model |> compile(
optimizer = optimizer_rmsprop(1e-3),
loss = list(
priority = loss_binary_crossentropy(from_logits = TRUE),
department = loss_categorical_crossentropy(from_logits = TRUE)
),
loss_weights = list(priority = 1.0, department = 0.2)
)
Train the model by passing lists of NumPy arrays of inputs and targets:
# Dummy input data
title_data <- random_integer(c(1280, 12), 0, num_words)
body_data <- random_integer(c(1280, 100), 0, num_words)
tags_data <- random_integer(c(1280, num_tags), 0, 2)
# Dummy target data
priority_targets <- random_normal(c(1280, 1))
dept_targets <- random_integer(c(1280, num_departments), 0, 2)
model |> fit(
list(title = title_data, body = body_data, tags = tags_data),
list(priority = priority_targets, department = dept_targets),
epochs = 2,
batch_size = 32
)
## Epoch 1/2
## 40/40 - 3s - 66ms/step - department_loss: 381.0319 - loss: 498.7886 - priority_loss: 117.7568
## Epoch 2/2
## 40/40 - 0s - 10ms/step - department_loss: 358.5705 - loss: 438.4252 - priority_loss: 79.8546
When calling fit with a Dataset
object, it should yield
either a list of lists like
list(list(title_data, body_data, tags_data), list(priority_targets, dept_targets))
or a list of named lists like
list(list(title = title_data, body = body_data, tags = tags_data), list(priority = priority_targets, department = dept_targets))
.
For more detailed explanation, refer to the training and evaluation guide.
In addition to models with multiple inputs and outputs, the
functional API makes it easy to manipulate non-linear connectivity
topologies – these are models with layers that are not connected
sequentially, which the Sequential
API cannot handle.
A common use case for this is residual connections. Let’s build a toy ResNet model for CIFAR10 to demonstrate this:
inputs <- keras_input(shape = c(32, 32, 3), name = "img")
block_1_output <- inputs |>
layer_conv_2d(32, kernel_size = 3, activation = "relu") |>
layer_conv_2d(64, kernel_size = 3, activation = "relu") |>
layer_max_pooling_2d(pool_size = 3)
block_2_output <- block_1_output |>
layer_conv_2d(32, kernel_size = 3, activation = "relu", padding = "same") |>
layer_conv_2d(64, kernel_size = 3, activation = "relu", padding = "same") |>
layer_add(block_1_output)
block_3_output <- block_2_output |>
layer_conv_2d(64, kernel_size = 3, activation = "relu", padding = "same") |>
layer_conv_2d(64, kernel_size = 3, activation = "relu", padding = "same") |>
layer_add(block_2_output)
outputs <- block_3_output |>
layer_conv_2d(64, 3, activation = "relu") |>
layer_global_average_pooling_2d() |>
layer_dense(256, activation = "relu") |>
layer_dropout(0.5) |>
layer_dense(10)
model <- keras_model(inputs, outputs, name = "toy_resnet")
summary(model)
## [1mModel: "toy_resnet"[0m
## ┏━━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━┓
## ┃[1m [0m[1mLayer (type) [0m[1m [0m┃[1m [0m[1mOutput Shape [0m[1m [0m┃[1m [0m[1m Param #[0m[1m [0m┃[1m [0m[1mConnected to [0m[1m [0m┃
## ┡━━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━┩
## │ img ([38;5;33mInputLayer[0m) │ ([38;5;45mNone[0m, [38;5;34m32[0m, [38;5;34m32[0m, [38;5;34m3[0m) │ [38;5;34m0[0m │ - │
## ├─────────────────────┼───────────────────┼────────────┼───────────────────┤
## │ conv2d_8 ([38;5;33mConv2D[0m) │ ([38;5;45mNone[0m, [38;5;34m30[0m, [38;5;34m30[0m, │ [38;5;34m896[0m │ img[[38;5;34m0[0m][[38;5;34m0[0m] │
## │ │ [38;5;34m32[0m) │ │ │
## ├─────────────────────┼───────────────────┼────────────┼───────────────────┤
## │ conv2d_9 ([38;5;33mConv2D[0m) │ ([38;5;45mNone[0m, [38;5;34m28[0m, [38;5;34m28[0m, │ [38;5;34m18,496[0m │ conv2d_8[[38;5;34m0[0m][[38;5;34m0[0m] │
## │ │ [38;5;34m64[0m) │ │ │
## ├─────────────────────┼───────────────────┼────────────┼───────────────────┤
## │ max_pooling2d_2 │ ([38;5;45mNone[0m, [38;5;34m9[0m, [38;5;34m9[0m, [38;5;34m64[0m) │ [38;5;34m0[0m │ conv2d_9[[38;5;34m0[0m][[38;5;34m0[0m] │
## │ ([38;5;33mMaxPooling2D[0m) │ │ │ │
## ├─────────────────────┼───────────────────┼────────────┼───────────────────┤
## │ conv2d_10 ([38;5;33mConv2D[0m) │ ([38;5;45mNone[0m, [38;5;34m9[0m, [38;5;34m9[0m, [38;5;34m32[0m) │ [38;5;34m18,464[0m │ max_pooling2d_2[[38;5;34m…[0m │
## ├─────────────────────┼───────────────────┼────────────┼───────────────────┤
## │ conv2d_11 ([38;5;33mConv2D[0m) │ ([38;5;45mNone[0m, [38;5;34m9[0m, [38;5;34m9[0m, [38;5;34m64[0m) │ [38;5;34m18,496[0m │ conv2d_10[[38;5;34m0[0m][[38;5;34m0[0m] │
## ├─────────────────────┼───────────────────┼────────────┼───────────────────┤
## │ add ([38;5;33mAdd[0m) │ ([38;5;45mNone[0m, [38;5;34m9[0m, [38;5;34m9[0m, [38;5;34m64[0m) │ [38;5;34m0[0m │ conv2d_11[[38;5;34m0[0m][[38;5;34m0[0m], │
## │ │ │ │ max_pooling2d_2[[38;5;34m…[0m │
## ├─────────────────────┼───────────────────┼────────────┼───────────────────┤
## │ conv2d_12 ([38;5;33mConv2D[0m) │ ([38;5;45mNone[0m, [38;5;34m9[0m, [38;5;34m9[0m, [38;5;34m64[0m) │ [38;5;34m36,928[0m │ add[[38;5;34m0[0m][[38;5;34m0[0m] │
## ├─────────────────────┼───────────────────┼────────────┼───────────────────┤
## │ conv2d_13 ([38;5;33mConv2D[0m) │ ([38;5;45mNone[0m, [38;5;34m9[0m, [38;5;34m9[0m, [38;5;34m64[0m) │ [38;5;34m36,928[0m │ conv2d_12[[38;5;34m0[0m][[38;5;34m0[0m] │
## ├─────────────────────┼───────────────────┼────────────┼───────────────────┤
## │ add_1 ([38;5;33mAdd[0m) │ ([38;5;45mNone[0m, [38;5;34m9[0m, [38;5;34m9[0m, [38;5;34m64[0m) │ [38;5;34m0[0m │ conv2d_13[[38;5;34m0[0m][[38;5;34m0[0m], │
## │ │ │ │ add[[38;5;34m0[0m][[38;5;34m0[0m] │
## ├─────────────────────┼───────────────────┼────────────┼───────────────────┤
## │ conv2d_14 ([38;5;33mConv2D[0m) │ ([38;5;45mNone[0m, [38;5;34m7[0m, [38;5;34m7[0m, [38;5;34m64[0m) │ [38;5;34m36,928[0m │ add_1[[38;5;34m0[0m][[38;5;34m0[0m] │
## ├─────────────────────┼───────────────────┼────────────┼───────────────────┤
## │ global_average_poo… │ ([38;5;45mNone[0m, [38;5;34m64[0m) │ [38;5;34m0[0m │ conv2d_14[[38;5;34m0[0m][[38;5;34m0[0m] │
## │ ([38;5;33mGlobalAveragePool…[0m │ │ │ │
## ├─────────────────────┼───────────────────┼────────────┼───────────────────┤
## │ dense_6 ([38;5;33mDense[0m) │ ([38;5;45mNone[0m, [38;5;34m256[0m) │ [38;5;34m16,640[0m │ global_average_p… │
## ├─────────────────────┼───────────────────┼────────────┼───────────────────┤
## │ dropout ([38;5;33mDropout[0m) │ ([38;5;45mNone[0m, [38;5;34m256[0m) │ [38;5;34m0[0m │ dense_6[[38;5;34m0[0m][[38;5;34m0[0m] │
## ├─────────────────────┼───────────────────┼────────────┼───────────────────┤
## │ dense_7 ([38;5;33mDense[0m) │ ([38;5;45mNone[0m, [38;5;34m10[0m) │ [38;5;34m2,570[0m │ dropout[[38;5;34m0[0m][[38;5;34m0[0m] │
## └─────────────────────┴───────────────────┴────────────┴───────────────────┘
## [1m Total params: [0m[38;5;34m186,346[0m (727.91 KB)
## [1m Trainable params: [0m[38;5;34m186,346[0m (727.91 KB)
## [1m Non-trainable params: [0m[38;5;34m0[0m (0.00 B)
Plot the model:
Now train the model:
c(c(x_train, y_train), c(x_test, y_test)) %<-% dataset_cifar10()
x_train <- x_train / 255.0
x_test <- x_test / 255.0
model |> compile(
optimizer = optimizer_rmsprop(1e-3),
loss = loss_sparse_categorical_crossentropy(from_logits = TRUE),
metrics = "acc"
)
# We restrict the data to the first 1000 samples so as to limit the
# guide render time.
# Try to train on the entire dataset until convergence!
model |> fit(
x_train[1:1000, , , ],
y_train[1:1000, ],
batch_size = 64,
epochs = 1,
validation_split = 0.2
)
## 13/13 - 5s - 375ms/step - acc: 0.1250 - loss: 2.3001 - val_acc: 0.1400 - val_loss: 2.2938
Because the graph of layers you are manipulating is a static data structure, it can be accessed and inspected. And this is how you are able to plot functional models as images.
This also means that you can access the activations of intermediate layers (“nodes” in the graph) and reuse them elsewhere – which is very useful for something like feature extraction.
Let’s look at an example. This is a VGG19 model with weights pretrained on ImageNet:
And these are the intermediate activations of the model, obtained by querying the graph data structure:
Use these features to create a new feature-extraction model that returns the values of the intermediate layer activations:
feat_extraction_model <- keras_model(inputs = vgg19$input,
outputs = features_list)
img <- random_normal(c(1, 224, 224, 3))
extracted_features <- feat_extraction_model(img)
This comes in handy for tasks like neural style transfer, among other things.
keras
includes a wide range of built-in layers, for
example:
conv_1d
, conv_2d
,
conv_3d
, conv_2d_transpose
max_pooling_1d
,
max_pooling_2d
, max_pooling_3d
,
average_pooling_3d
gru
, lstm
,
conv_lstm_2d
batch_normalization
, dropout
,
embedding
, etc.But if you don’t find what you need, it’s easy to extend the API by
creating your own layers. All layers subclass the Layer
class and implement:
call
method, that specifies the computation done by the
layer.build
method, that creates the weights of the layer
(this is just a style convention since you can create weights in
initialize
, as well).To learn more about creating layers from scratch, read custom layers and models guide.
The following is a basic implementation of
layer_dense()
:
custom_dense <- Layer(
classname = "CustomDense",
initialize = function(units = 32) {
super$initialize()
self$units <- as.integer(units)
},
build = function(input_shape) {
self$w <- self$add_weight(
shape = shape(input_shape[[2]], self$units),
initializer = "random_normal",
trainable = TRUE,
)
self$b <- self$add_weight(
shape = shape(self$units),
initializer="random_normal",
trainable = TRUE
)
},
call = function(inputs) {
op_matmul(inputs, self$w) + self$b
}
)
inputs <- keras_input(c(4))
outputs <- custom_dense(inputs, 10)
model <- keras_model(inputs, outputs)
For serialization support in your custom layer, define a
get_config()
method that returns the constructor arguments
of the layer instance:
custom_dense <- Layer(
classname = "CustomDense",
initialize = function(units = 32, ...) {
super$initialize()
self$units <- as.integer(units)
},
build = function(input_shape) {
self$w <- self$add_weight(
shape = shape(input_shape[[2]], self$units),
initializer = "random_normal",
trainable = TRUE,
)
self$b <- self$add_weight(
shape = shape(self$units),
initializer="random_normal",
trainable = TRUE
)
},
call = function(inputs) {
op_matmul(inputs, self$w) + self$b
},
get_config = function() {
list(units = self$units)
}
)
inputs <- keras_input(c(4))
outputs <- custom_dense(inputs, 10)
model <- keras_model(inputs, outputs)
config <- get_config(model)
new_model <-
from_config(config, custom_objects = list(CustomDense = custom_dense))
Optionally, implement the class method
from_config(cls, config)
which is used when recreating a
layer instance given its config dictionary. The default implementation
of from_config
is:
Should you use the Keras functional API to create a new model, or
just subclass the Model
class directly? In general, the
functional API is higher-level, easier and safer, and has a number of
features that subclassed models do not support.
However, model subclassing provides greater flexibility when building
models that are not easily expressible as directed acyclic graphs of
layers. For example, you could not implement a Tree-RNN with the
functional API and would have to subclass Model
directly.
For an in-depth look at the differences between the functional API and model subclassing, read What are Symbolic and Imperative APIs in TensorFlow 2.0?.
The following properties are also true for Sequential models (which are also data structures), but are not true for subclassed models (which are R and Python (byte)code, not data structures).
There is no super$initialize(...)
, no
call = function(...)
, no self$...
, etc.
Compare:
inputs <- keras_input(shape = shape(32))
outputs <- inputs |>
layer_dense(64, activation = "relu") |>
layer_dense(10)
mlp <- keras_model(inputs, outputs)
With the subclassed version:
MLP <- Model(
classname = "MLP",
initialize = function(...) {
super$initialize(...)
self$dense_1 <- layer_dense(units = 64, activation = "relu")
self$dense_2 <- layer_dense(units = 10)
},
call = function(inputs) {
inputs |>
self$dense_1() |>
self$dense_2()
}
)
# Instantiate the model.
mlp <- MLP()
# Necessary to create the model's state.
# The model doesn't have a state until it's called at least once.
out <- mlp(op_zeros(c(1, 32)))
In the functional API, the input specification (shape and dtype) is
created in advance (using Input
). Every time you call a
layer, the layer checks that the specification passed to it matches its
assumptions, and it will raise a helpful error message if not.
This guarantees that any model you can build with the functional API will run. All debugging – other than convergence-related debugging – happens statically during the model construction and not at execution time. This is similar to type checking in a compiler.
You can plot the model as a graph, and you can easily access intermediate nodes in this graph. For example, to extract and reuse the activations of intermediate layers (as seen in a previous example):
Because a functional model is a data structure rather than a piece of code, it is safely serializable and can be saved as a single file that allows you to recreate the exact same model without having access to any of the original code. See the serialization & saving guide.
To serialize a subclassed model, it is necessary for the implementer
to specify a get_config()
and from_config()
method at the model level.
The functional API treats models as DAGs of layers. This is true for most deep learning architectures, but not all – for example, recursive networks or Tree RNNs do not follow this assumption and cannot be implemented in the functional API.
Choosing between the functional API or Model subclassing isn’t a
binary decision that restricts you into one category of models. All
models in the keras
API can interact with each other,
whether they’re Sequential
models, functional models, or
subclassed models that are written from scratch.
You can always use a functional model or Sequential
model as part of a subclassed model or layer:
units <- 32
timesteps <- 10
input_dim <- 5
# Define a Functional model
inputs <- keras_input(shape(NA, units))
outputs <- inputs |>
layer_global_average_pooling_1d() |>
layer_dense(units = 1)
model <- keras_model(inputs, outputs)
layer_custom_rnn <- Layer(
classname = "CustomRNN",
initialize = function(...) {
super$initialize(...)
self$units <- units
self$projection_1 <- layer_dense(units = units, activation = "tanh")
self$projection_2 <- layer_dense(units = units, activation = "tanh")
self$classifier <- model
},
call = function(inputs, ...) {
outputs <- list()
state <- op_zeros(c(shape(inputs)[[1]], self$units))
for (t in 1:(shape(inputs)[[2]])) {
x <- inputs[, t, ]
h <- self$projection_1(x)
y <- h + self$projection_2(state)
state <- y
outputs[[t]] <- y
}
features <- op_stack(outputs, axis = 2)
self$classifier(features)
}
)
rnn <- layer_custom_rnn()
out <- rnn(op_zeros(c(1, timesteps, input_dim)))
You can use any subclassed layer or model in the functional API as
long as it implements a call
method that follows one of the
following patterns:
call(inputs, ...)
– Where inputs
is a
tensor or a nested structure of tensors (e.g. a list of tensors), and
where ...
are non-tensor arguments (non-inputs).call(inputs, training = NULL, ...)
– Where
training
is a boolean indicating whether the layer should
behave in training mode and inference mode.call(inputs, mask = NULL, ...)
– Where
mask
is a boolean mask tensor (useful for RNNs, for
instance).call(inputs, training = NULL, mask = NULL, ...)
– Of
course, you can have both masking and training-specific behavior at the
same time.Additionally, if you implement the get_config()
method
on your custom Layer or model, the functional models you create will
still be serializable and cloneable.
Here’s a quick example of a custom RNN, written from scratch, being used in a functional model:
units <- 32
timesteps <- 10
input_dim <- 5
batch_size <- 16
layer_custom_rnn <- Layer(
"custom_rnn",
initialize = function(...) {
super$initialize(...)
self$units <- units
self$projection_1 <- layer_dense(units = units, activation = "tanh")
self$projection_2 <- layer_dense(units = units, activation = "tanh")
self$classifier <- layer_dense(units = 1)
},
call = function(inputs, ...) {
outputs <- list()
state <- op_zeros(c(shape(inputs)[[1]], self$units))
for (t in 1:(shape(inputs)[[2]])) {
x <- inputs[, t, ]
h <- self$projection_1(x)
y <- h + self$projection_2(state)
state <- y
outputs[[t]] <- y
}
features <- op_stack(outputs, axis = 2)
self$classifier(features)
}
)
# Note that you specify a static batch size for the inputs with the `batch_shape`
# arg, because the inner computation of `layer_custom_rnn()` requires a static batch size
# (when you create the `state` zeros tensor).
inputs <- keras_input(batch_shape = shape(batch_size, timesteps, input_dim))
outputs <- inputs |>
layer_conv_1d(filters = 32, kernel_size = 3) |>
layer_custom_rnn()
model <- keras_model(inputs, outputs)
out <- model(op_zeros(c(1, 10, 5)))