Test it on the Peltarion Platform

A platform to build and deploy deep learning projects.

Even if you’re not an AI superstar.

TRY FOR FREE

Dense

The dense block represents a fully connected layer of artificial nodes.

Each of the nodes (as many as the Nodes attribute) has one weight per input feature and an output that is a function of its inputs.

2 Dense blocks
Figure 1. Two Dense blocks

Usage of Dense blocks

Last block before the Target block

In a classification problem, you should add a Dense block as the last block before the Target block. Set as many nodes as the number of classes the model can predict.

Flatten the input to a Dense blocks

If the input has more than one dimension of feature, e.g., an image has 3 (height, width, channels), the data should be flattened before being fed into the nodes with a Flatten block.

Knowledge stored in the Dense block

The more Dense blocks you add, the more complex function approximation can be done by the net.

Dense blocks are the only blocks with nodes used in multilayer perceptrons, the simplest form of deep neural networks.

Dense blocks with many channels grow quickly in memory usage

A Dense block, where all nodes are fully connected, and each feature is treated separately, grow quickly in memory usage with the input size. This means that Dense blocks aren’t suitable for data like images where the input is height x width x number of channels, which for a standard definition image is at least 1 million.
Convolutional blocks, e.g., the 2D Convolution block, take better advantage of the structure within image data and need less memory to work.

Parameters

Nodes: the number of nodes in the block.

Initializer: the procedure used to set the initial values of the weights and the bias before starting training.
Default: Glorot uniform initialization

Activation: the function used to transform the output of the dot product inside the block.
Default: ReLU

Trainable: whether we want the training algorithm to change the value of the weights during training. In some cases, one will want to keep parts of the network static, e.g., when using the encoder part of an autoencoder as preprocessing for another model.