Writing your own Keras layers
For simple, stateless custom operations, you are probably better off using
layers.core.Lambda layers. But for any custom operation that has trainable weights, you should implement your own layer.
Here is the skeleton of a Keras layer, as of Keras 1.1.3 (if you have an older version, please upgrade). There are only three methods you need to implement:
build(input_shape): this is where you will define your weights. This method must set
self.built = True, which can be done by calling
call(x): this is where the layer's logic lives. Unless you want your layer to support masking, you only have to care about the first argument passed to
call: the input tensor.
get_output_shape_for(input_shape): in case your layer modifies the shape of its input, you should specify here the shape transformation logic. This allows Keras to do automatic shape inference.
from keras import backend as K from keras.engine.topology import Layer import numpy as np class MyLayer(Layer): def __init__(self, output_dim, **kwargs): self.output_dim = output_dim super(MyLayer, self).__init__(**kwargs) def build(self, input_shape): # Create a trainable weight variable for this layer. self.W = self.add_weight(shape=(input_shape, self.output_dim), initializer='random_uniform', trainable=True) super(MyLayer, self).build() # Be sure to call this somewhere! def call(self, x, mask=None): return K.dot(x, self.W) def get_output_shape_for(self, input_shape): return (input_shape, self.output_dim)
The existing Keras layers provide ample examples of how to implement almost anything. Never hesitate to read the source code!