|
Java example source code file (MultiLayerNetwork.java)
This example Java source code file (MultiLayerNetwork.java) is included in the alvinalexander.com
"Java Source Code
Warehouse" project. The intent of this project is to help you "Learn
Java by Example" TM.
Learn more about this Java project at its project page.
The MultiLayerNetwork.java Java example source code
/*
*
* * Copyright 2015 Skymind,Inc.
* *
* * Licensed under the Apache License, Version 2.0 (the "License");
* * you may not use this file except in compliance with the License.
* * You may obtain a copy of the License at
* *
* * http://www.apache.org/licenses/LICENSE-2.0
* *
* * Unless required by applicable law or agreed to in writing, software
* * distributed under the License is distributed on an "AS IS" BASIS,
* * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* * See the License for the specific language governing permissions and
* * limitations under the License.
*
*/
package org.deeplearning4j.nn.multilayer;
import lombok.Setter;
import org.deeplearning4j.berkeley.Pair;
import org.deeplearning4j.berkeley.Triple;
import org.deeplearning4j.datasets.iterator.AsyncDataSetIterator;
import org.deeplearning4j.datasets.iterator.MultipleEpochsIterator;
import org.deeplearning4j.datasets.iterator.impl.ListDataSetIterator;
import org.deeplearning4j.eval.Evaluation;
import org.deeplearning4j.nn.api.*;
import org.deeplearning4j.nn.conf.BackpropType;
import org.deeplearning4j.nn.conf.MultiLayerConfiguration;
import org.deeplearning4j.nn.conf.NeuralNetConfiguration;
import org.deeplearning4j.nn.gradient.DefaultGradient;
import org.deeplearning4j.nn.gradient.Gradient;
import org.deeplearning4j.nn.layers.BaseOutputLayer;
import org.deeplearning4j.nn.layers.BasePretrainNetwork;
import org.deeplearning4j.nn.layers.factory.LayerFactories;
import org.deeplearning4j.nn.layers.recurrent.BaseRecurrentLayer;
import org.deeplearning4j.nn.params.DefaultParamInitializer;
import org.deeplearning4j.nn.updater.UpdaterCreator;
import org.deeplearning4j.nn.weights.WeightInit;
import org.deeplearning4j.optimize.Solver;
import org.deeplearning4j.optimize.api.ConvexOptimizer;
import org.deeplearning4j.optimize.api.IterationListener;
import org.deeplearning4j.util.ModelSerializer;
import org.deeplearning4j.util.MultiLayerUtil;
import org.deeplearning4j.util.TimeSeriesUtils;
import org.nd4j.linalg.api.ndarray.INDArray;
import org.nd4j.linalg.dataset.DataSet;
import org.nd4j.linalg.dataset.api.iterator.DataSetIterator;
import org.nd4j.linalg.factory.Nd4j;
import org.nd4j.linalg.heartbeat.Heartbeat;
import org.nd4j.linalg.heartbeat.reports.Environment;
import org.nd4j.linalg.heartbeat.reports.Event;
import org.nd4j.linalg.heartbeat.reports.Task;
import org.nd4j.linalg.heartbeat.utils.EnvironmentUtils;
import org.nd4j.linalg.heartbeat.utils.TaskUtils;
import org.nd4j.linalg.indexing.NDArrayIndex;
import org.nd4j.linalg.ops.transforms.Transforms;
import org.nd4j.linalg.util.FeatureUtil;
import org.nd4j.linalg.util.LinAlgExceptions;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import java.io.Serializable;
import java.lang.String;
import java.lang.reflect.Constructor;
import java.util.*;
/**
* A base class for a multi
* layer neural network with a logistic output layer
* and multiple hidden neuralNets.
*
* @author Adam Gibson
*/
public class MultiLayerNetwork implements Serializable, Classifier, Layer {
private static final Logger log = LoggerFactory.getLogger(MultiLayerNetwork.class);
//the hidden neuralNets
protected Layer[] layers;
protected LinkedHashMap<String, Layer> layerMap = new LinkedHashMap<>();
//default training examples and associated neuralNets
protected INDArray input, labels;
//sometimes we may need to transform weights; this allows a
protected boolean initCalled = false;
private Collection<IterationListener> listeners = new ArrayList<>();
protected NeuralNetConfiguration defaultConfiguration;
protected MultiLayerConfiguration layerWiseConfigurations;
protected Gradient gradient;
protected INDArray epsilon;
protected double score;
@Setter protected boolean initDone = false;
protected INDArray flattenedParams; //Params for all layers are a view/subset of this array
protected transient INDArray flattenedGradients; //Gradients for all layers are a view/subset of this array
/*
Binary drop connect mask
*/
protected INDArray mask;
protected int layerIndex; //For Layer.get/setIndex()
protected transient Solver solver; //Used to call optimizers during backprop
public MultiLayerNetwork(MultiLayerConfiguration conf) {
this.layerWiseConfigurations = conf;
this.defaultConfiguration = conf.getConf(0).clone();
}
/**
* Initialize the network based on the configuration
*
* @param conf the configuration json
* @param params the parameters
*/
public MultiLayerNetwork(String conf, INDArray params) {
this(MultiLayerConfiguration.fromJson(conf));
init();
setParameters(params);
}
/**
* Initialize the network based on the configuraiton
*
* @param conf the configuration
* @param params the parameters
*/
public MultiLayerNetwork(MultiLayerConfiguration conf, INDArray params) {
this(conf);
init();
setParameters(params);
}
protected void intializeConfigurations() {
if (layerWiseConfigurations == null)
layerWiseConfigurations = new MultiLayerConfiguration.Builder().build();
if (layers == null)
layers = new Layer[getnLayers()];
if (defaultConfiguration == null)
defaultConfiguration = new NeuralNetConfiguration.Builder()
.build();
}
/**
* This unsupervised learning method runs
* contrastive divergence on each RBM layer in the network.
*
* @param iter the input to iterate on
* The typical tip is that the higher k is the closer to the model
* you will be approximating due to more sampling. K = 1
* usually gives very good results and is the default in quite a few situations.
*/
public void pretrain(DataSetIterator iter) {
if (!layerWiseConfigurations.isPretrain())
return;
INDArray layerInput;
for (int i = 0; i < getnLayers(); i++) {
if (i == 0) {
while (iter.hasNext()) {
DataSet next = iter.next();
if(getLayerWiseConfigurations().getInputPreProcess(i) != null) {
INDArray features = next.getFeatureMatrix();
layerInput = getLayerWiseConfigurations().getInputPreProcess(i).preProcess(features, features.size(0));
}
else
layerInput = next.getFeatureMatrix();
setInput(layerInput);
/*During pretrain, feed forward expected activations of network, use activation cooccurrences during pretrain */
if (this.getInput() == null || this.getLayers() == null)
initializeLayers(input());
layers[i].fit(input());
log.info("Training on layer " + (i + 1) + " with " + input().size(0) + " examples");
}
} else {
while (iter.hasNext()) {
DataSet next = iter.next();
layerInput = next.getFeatureMatrix();
for (int j = 1; j <= i; j++)
layerInput = activationFromPrevLayer(j - 1, layerInput,true);
log.info("Training on layer " + (i + 1) + " with " + layerInput.size(0) + " examples");
getLayer(i).fit(layerInput);
}
}
iter.reset();
}
}
/**
* This unsupervised learning method runs
* contrastive divergence on each RBM layer in the network.
*
* @param input the input to iterate on
* The typical tip is that the higher k is the closer to the model
* you will be approximating due to more sampling. K = 1
* usually gives very good results and is the default in quite a few situations.
*/
public void pretrain(INDArray input) {
if (!layerWiseConfigurations.isPretrain())
return;
/* During pretrain, feed forward expected activations of network, use activation cooccurrences during pretrain */
int miniBatchSize = input.size(0);
INDArray layerInput = null;
for (int i = 0; i < getnLayers() - 1; i++) {
if (i == 0)
if(getLayerWiseConfigurations().getInputPreProcess(i) != null)
layerInput = getLayerWiseConfigurations().getInputPreProcess(i).preProcess(input,miniBatchSize);
else
layerInput = input;
else
layerInput = activationFromPrevLayer(i - 1, layerInput,true);
log.info("Training on layer " + (i + 1) + " with " + layerInput.size(0) + " examples");
getLayers()[i].fit(layerInput);
}
}
@Override
public int batchSize() {
return input.size(0);
}
@Override
public NeuralNetConfiguration conf() {
return defaultConfiguration;
}
@Override
public void setConf(NeuralNetConfiguration conf) {
throw new UnsupportedOperationException();
}
@Override
public INDArray input() {
return input;
}
@Override
public void validateInput() {
}
@Override
public ConvexOptimizer getOptimizer() {
throw new UnsupportedOperationException();
}
@Override
public INDArray getParam(String param) {
//Get params for MultiLayerNetwork sub layers.
//Parameter keys here: same as MultiLayerNetwork.backprop().
int idx = param.indexOf("_");
if( idx == -1 ) throw new IllegalStateException("Invalid param key: not have layer separator: \""+param+"\"");
int layerIdx = Integer.parseInt(param.substring(0, idx));
String newKey = param.substring(idx+1);
return layers[layerIdx].getParam(newKey);
}
@Override
public void initParams() {
throw new UnsupportedOperationException();
}
@Override
public Map<String, INDArray> paramTable() {
//Get all parameters from all layers
Map<String,INDArray> allParams = new LinkedHashMap<>();
for( int i=0; i<layers.length; i++ ){
Map<String,INDArray> paramMap = layers[i].paramTable();
for( Map.Entry<String, INDArray> entry : paramMap.entrySet() ){
String newKey = i + "_" + entry.getKey();
allParams.put(newKey, entry.getValue());
}
}
return allParams;
}
@Override
public void setParamTable(Map<String, INDArray> paramTable) {
throw new UnsupportedOperationException();
}
@Override
public void setParam(String key, INDArray val) {
//Set params for MultiLayerNetwork sub layers.
//Parameter keys here: same as MultiLayerNetwork.backprop().
int idx = key.indexOf("_");
if( idx == -1 ) throw new IllegalStateException("Invalid param key: not have layer separator: \""+key+"\"");
int layerIdx = Integer.parseInt(key.substring(0, idx));
String newKey = key.substring(idx+1);
layers[layerIdx].setParam(newKey,val);
}
public MultiLayerConfiguration getLayerWiseConfigurations() {
return layerWiseConfigurations;
}
public void setLayerWiseConfigurations(MultiLayerConfiguration layerWiseConfigurations) {
this.layerWiseConfigurations = layerWiseConfigurations;
}
/**
* Base class for initializing the neuralNets based on the input.
* This is meant for capturing numbers such as input columns or other things.
*
* @param input the input matrix for training
*/
public void initializeLayers(INDArray input) {
if (input == null)
throw new IllegalArgumentException("Unable to initialize neuralNets with empty input");
this.input = input;
setInputMiniBatchSize(input.size(0));
if (!initCalled)
init();
}
/**
* Initialize the MultiLayerNetwork. This should be called once before the network is used.
*/
public void init() {
init(null,false);
}
/**
* Initialize the MultiLayerNetwork, optionally with an existing parameters array.
* If an existing parameters array is specified, it will be used (and the values will not be modified) in the network;
* if no parameters array is specified, parameters will be initialized randomly according to the network configuration.
*
* @param parameters Network parameter. May be null. If null: randomly initialize.
* @param cloneParametersArray Whether the parameter array (if any) should be cloned, or used directly
*/
public void init(INDArray parameters, boolean cloneParametersArray){
if (layerWiseConfigurations == null || layers == null)
intializeConfigurations();
if (initCalled)
return;
int nLayers = getnLayers();
if (nLayers < 1)
throw new IllegalStateException("Unable to create network: number of layers is less than 1");
if (this.layers == null || this.layers[0] == null) {
if (this.layers == null)
this.layers = new Layer[nLayers];
//First: Work out total length of (backprop) params
int backpropParamLength = 0;
int[] nParamsPerLayer = new int[nLayers];
for( int i=0; i<nLayers; i++ ){
NeuralNetConfiguration conf = layerWiseConfigurations.getConf(i);
nParamsPerLayer[i] = LayerFactories.getFactory(conf).initializer().numParams(conf,true);
backpropParamLength += nParamsPerLayer[i];
}
//Create parameters array, if required
boolean initializeParams;
if(parameters != null ){
if(!parameters.isRowVector()) throw new IllegalArgumentException("Invalid parameters: should be a row vector");
if(parameters.length() != backpropParamLength) throw new IllegalArgumentException("Invalid parameters: expected length " + backpropParamLength + ", got length " + parameters.length());
if(cloneParametersArray) flattenedParams = parameters.dup();
else flattenedParams = parameters;
initializeParams = false;
} else {
flattenedParams = Nd4j.create(1, backpropParamLength);
initializeParams = true;
}
// construct multi-layer
int paramCountSoFar = 0;
for (int i = 0; i < nLayers; i++) {
INDArray paramsView;
if(nParamsPerLayer[i] > 0){
paramsView = flattenedParams.get(NDArrayIndex.point(0), NDArrayIndex.interval(paramCountSoFar, paramCountSoFar + nParamsPerLayer[i]));
} else {
paramsView = null;
}
paramCountSoFar += nParamsPerLayer[i];
NeuralNetConfiguration conf = layerWiseConfigurations.getConf(i);
layers[i] = LayerFactories.getFactory(conf).create(conf, listeners, i, paramsView, initializeParams);
layerMap.put(conf.getLayer().getLayerName(), layers[i]);
}
initCalled = true;
initMask();
}
//Set parameters in MultiLayerNetwork.defaultConfiguration for later use in BaseOptimizer.setupSearchState() etc
//Keyed as per backprop()
defaultConfiguration.clearVariables();
for( int i=0; i<layers.length; i++ ){
for( String s : layers[i].conf().variables() ){
defaultConfiguration.addVariable(i+"_"+s);
}
}
}
public boolean isInitCalled(){
return initCalled;
}
/**
* This method: initializes the flattened gradients array (used in backprop) and sets the appropriate subset in all layers.
*/
protected void initGradientsView(){
if(layers == null) init();
int nLayers = layers.length;
//First: Work out total length of (backprop) params
int backpropParamLength = 0;
int[] nParamsPerLayer = new int[nLayers];
for( int i=0; i<nLayers; i++ ){
NeuralNetConfiguration conf = layerWiseConfigurations.getConf(i);
nParamsPerLayer[i] = LayerFactories.getFactory(conf).initializer().numParams(conf,true);
backpropParamLength += nParamsPerLayer[i];
}
flattenedGradients = Nd4j.createUninitialized(new int[]{1,backpropParamLength},'f'); //No need to initialize, as each layer will do it each iteration anyway
int backpropParamsSoFar = 0;
for(int i=0; i<layers.length; i++ ){
if(nParamsPerLayer[i] == 0) continue; //This layer doesn't have any parameters...
INDArray thisLayerGradView = flattenedGradients.get(NDArrayIndex.point(0), NDArrayIndex.interval(backpropParamsSoFar, backpropParamsSoFar + nParamsPerLayer[i]));
layers[i].setBackpropGradientsViewArray(thisLayerGradView);
backpropParamsSoFar += nParamsPerLayer[i];
}
}
/**
* Triggers the activation of the last hidden layer ie: not logistic regression
*
* @return the activation of the last hidden layer given the last input to the network
*/
public INDArray activate() {
return getLayers()[getLayers().length - 1].activate();
}
/**
* Triggers the activation for a given layer
*
* @param layer the layer to activate on
* @return the activation for a given layer
*/
public INDArray activate(int layer) {
return getLayer(layer).activate();
}
@Override
public INDArray activate(INDArray input) {
throw new UnsupportedOperationException();
}
/**
* Triggers the activation of the given layer
*
* @param layer the layer to trigger on
* @param input the input to the hidden layer
* @return the activation of the layer based on the input
*/
public INDArray activate(int layer, INDArray input) {
return getLayer(layer).activate(input);
}
@Override
public INDArray activationMean() {
//TODO determine how to pass back all activationMean for MLN
throw new UnsupportedOperationException();
// List<INDArray> avgActivations = new ArrayList<>();
//
// for( Layer layer: getLayers() ){
// avgActivations.add(layer.activationMean());
// }
// return Nd4j.toFlattened(avgActivations);
}
/**
* Sets the input and labels from this dataset
*
* @param data the dataset to initialize with
*/
public void initialize(DataSet data) {
setInput(data.getFeatureMatrix());
feedForward(getInput());
this.labels = data.getLabels();
if (getOutputLayer() instanceof BaseOutputLayer) {
BaseOutputLayer o = (BaseOutputLayer) getOutputLayer();
o.setLabels(labels);
}
}
/**
* Compute input linear transformation (z) from previous layer
* Apply pre processing transformation where necessary
*
* @param curr the current layer
* @param input the input
* @param training training or test mode
* @return the activation from the previous layer
*/
public INDArray zFromPrevLayer(int curr, INDArray input,boolean training) {
if(getLayerWiseConfigurations().getInputPreProcess(curr) != null)
input = getLayerWiseConfigurations().getInputPreProcess(curr).preProcess(input,input.size(0));
INDArray ret = layers[curr].preOutput(input, training);
return ret;
}
/**
* Calculate activation from previous layer including pre processing where necessary
*
* @param curr the current layer
* @param input the input
* @return the activation from the previous layer
*/
public INDArray activationFromPrevLayer(int curr, INDArray input,boolean training) {
if(getLayerWiseConfigurations().getInputPreProcess(curr) != null)
input = getLayerWiseConfigurations().getInputPreProcess(curr).preProcess(input,getInputMiniBatchSize());
INDArray ret = layers[curr].activate(input, training);
return ret;
}
/**
* Calculate activation for few layers at once. Suitable for autoencoder partial activation.
*
* In example: in 10-layer deep autoencoder, layers 0 - 4 inclusive are used for encoding part, and layers 5-9 inclusive are used for decoding part.
*
* @param from first layer to be activated, inclusive
* @param to last layer to be activated, inclusive
* @return the activation from the last layer
*/
public INDArray activateSelectedLayers(int from, int to, INDArray input) {
if (input == null) throw new IllegalStateException("Unable to perform activation; no input found");
if (from < 0 || from >= layers.length || from >= to) throw new IllegalStateException("Unable to perform activation; FROM is out of layer space");
if (to < 1 || to >= layers.length) throw new IllegalStateException("Unable to perform activation; TO is out of layer space");
INDArray res = input;
for (int l = from; l <= to; l++) {
res = this.activationFromPrevLayer(l, res, false);
}
return res;
}
/**
* * Compute input linear transformation (z) of the output layer
*
* @return the list of activations for each layer
*/
public List<INDArray> computeZ(boolean training) {
INDArray currInput = this.input;
List<INDArray> activations = new ArrayList<>();
activations.add(currInput);
for (int i = 0; i < layers.length; i++) {
currInput = zFromPrevLayer(i, currInput,training);
//applies drop connect to the activation
activations.add(currInput);
}
return activations;
}
/**
* Compute activations from input to output of the output layer
*
* @return the list of activations for each layer
*/
public List<INDArray> computeZ(INDArray input,boolean training) {
if (input == null)
throw new IllegalStateException("Unable to perform feed forward; no input found");
else if (this.getLayerWiseConfigurations().getInputPreProcess(0) != null)
setInput(getLayerWiseConfigurations().getInputPreProcess(0).preProcess(input,getInputMiniBatchSize()));
else
setInput(input);
return computeZ(training);
}
/**
* Compute activations from input to output of the output layer
*
* @return the list of activations for each layer
*/
public List<INDArray> feedForward(INDArray input, boolean train) {
setInput(input);
return feedForward(train);
}
/**
* Compute activations from input to output of the output layer
*
* @return the list of activations for each layer
*/
public List<INDArray> feedForward(boolean train) {
return feedForwardToLayer(layers.length - 1, train);
}
/** Compute the activations from the input to the specified layer.<br>
* To compute activations for all layers, use feedForward(...) methods<br>
* Note: output list includes the original input. So list.get(0) is always the original input, and
* list.get(i+1) is the activations of the ith layer.
* @param layerNum Index of the last layer to calculate activations for. Layers are zero-indexed.
* feedForwardToLayer(i,input) will return the activations for layers 0..i (inclusive)
* @param input Input to the network
* @return list of activations.
*/
public List<INDArray> feedForwardToLayer(int layerNum, INDArray input){
return feedForwardToLayer(layerNum, input, false);
}
/** Compute the activations from the input to the specified layer.<br>
* To compute activations for all layers, use feedForward(...) methods<br>
* Note: output list includes the original input. So list.get(0) is always the original input, and
* list.get(i+1) is the activations of the ith layer.
* @param layerNum Index of the last layer to calculate activations for. Layers are zero-indexed.
* feedForwardToLayer(i,input) will return the activations for layers 0..i (inclusive)
* @param input Input to the network
* @param train true for training, false for test (i.e., false if using network after training)
* @return list of activations.
*/
public List<INDArray> feedForwardToLayer(int layerNum, INDArray input, boolean train){
setInput(input);
return feedForwardToLayer(layerNum, train);
}
/** Compute the activations from the input to the specified layer, using the currently set input for the network.<br>
* To compute activations for all layers, use feedForward(...) methods<br>
* Note: output list includes the original input. So list.get(0) is always the original input, and
* list.get(i+1) is the activations of the ith layer.
* @param layerNum Index of the last layer to calculate activations for. Layers are zero-indexed.
* feedForwardToLayer(i,input) will return the activations for layers 0..i (inclusive)
* @param train true for training, false for test (i.e., false if using network after training)
* @return list of activations.
*/
public List<INDArray> feedForwardToLayer(int layerNum, boolean train){
INDArray currInput = input;
List<INDArray> activations = new ArrayList<>();
activations.add(currInput);
for (int i = 0; i <= layerNum; i++) {
currInput = activationFromPrevLayer(i, currInput,train);
//applies drop connect to the activation
activations.add(currInput);
}
return activations;
}
/**
* Compute activations from input to output of the output layer
*
* @return the list of activations for each layer
*/
public List<INDArray> feedForward() {
return feedForward(false);
}
/**
* Compute activations from input to output of the output layer
*
* @return the list of activations for each layer
*/
public List<INDArray> feedForward(INDArray input) {
if (input == null)
throw new IllegalStateException("Unable to perform feed forward; no input found");
else if (this.getLayerWiseConfigurations().getInputPreProcess(0) != null)
setInput(getLayerWiseConfigurations().getInputPreProcess(0).preProcess(input,input.size(0)));
else
setInput(input);
return feedForward();
}
/** Compute the activations from the input to the output layer, given mask arrays (that may be null)
* The masking arrays are used in situations such an one-to-many and many-to-one rucerrent neural network (RNN)
* designs, as well as for supporting time series of varying lengths within the same minibatch for RNNs.
*/
public List<INDArray> feedForward(INDArray input, INDArray featuresMask, INDArray labelsMask){
setLayerMaskArrays(featuresMask,labelsMask);
List<INDArray> list = feedForward(input);
clearLayerMaskArrays();
return list;
}
@Override
public Gradient gradient() {
return gradient;
}
public INDArray epsilon() {
return epsilon;
}
@Override
public Pair<Gradient, Double> gradientAndScore() {
return new Pair<>(gradient(), score());
}
/* delta computation for back prop with the R operator */
protected List<INDArray> computeDeltasR(INDArray v) {
List<INDArray> deltaRet = new ArrayList<>();
INDArray[] deltas = new INDArray[getnLayers() + 1];
List<INDArray> activations = feedForward();
List<INDArray> rActivations = feedForwardR(activations, v);
/*
* Precompute activations and z's (pre activation network outputs)
*/
List<INDArray> weights = new ArrayList<>();
List<INDArray> biases = new ArrayList<>();
List<String> activationFunctions = new ArrayList<>();
for (int j = 0; j < getLayers().length; j++) {
weights.add(getLayers()[j].getParam(DefaultParamInitializer.WEIGHT_KEY));
biases.add(getLayers()[j].getParam(DefaultParamInitializer.BIAS_KEY));
activationFunctions.add(getLayers()[j].conf().getLayer().getActivationFunction());
}
INDArray rix = rActivations.get(rActivations.size() - 1).divi((double) input.size(0));
LinAlgExceptions.assertValidNum(rix);
//errors
for (int i = getnLayers() - 1; i >= 0; i--) {
//W^t * error^l + 1
deltas[i] = activations.get(i).transpose().mmul(rix);
if (i > 0)
rix = rix.mmul(weights.get(i).addRowVector(biases.get(i)).transpose()).muli(Nd4j.getExecutioner().execAndReturn(Nd4j.getOpFactory().createTransform(activationFunctions.get(i - 1), activations.get(i)).derivative()));
}
for (int i = 0; i < deltas.length - 1; i++) {
deltaRet.add(deltas[i]);
}
return deltaRet;
}
/* delta computation for back prop with precon for SFH */
protected List<Pair computeDeltas2() {
List<Pair deltaRet = new ArrayList<>();
List<INDArray> activations = feedForward();
INDArray[] deltas = new INDArray[activations.size() - 1];
INDArray[] preCons = new INDArray[activations.size() - 1];
//- y - h
INDArray ix = activations.get(activations.size() - 1).sub(labels).div(labels.size(0));
/*
* Precompute activations and z's (pre activation network outputs)
*/
List<INDArray> weights = new ArrayList<>();
List<INDArray> biases = new ArrayList<>();
List<String> activationFunctions = new ArrayList<>();
for (int j = 0; j < getLayers().length; j++) {
weights.add(getLayers()[j].getParam(DefaultParamInitializer.WEIGHT_KEY));
biases.add(getLayers()[j].getParam(DefaultParamInitializer.BIAS_KEY));
activationFunctions.add(getLayers()[j].conf().getLayer().getActivationFunction());
}
//errors
for (int i = weights.size() - 1; i >= 0; i--) {
deltas[i] = activations.get(i).transpose().mmul(ix);
preCons[i] = Transforms.pow(activations.get(i).transpose(), 2).mmul(Transforms.pow(ix, 2)).muli(labels.size(0));
if (i > 0) {
//W[i] + b[i] * f'(z[i - 1])
ix = ix.mmul(weights.get(i).transpose()).muli(Nd4j.getExecutioner().execAndReturn(Nd4j.getOpFactory().createTransform(activationFunctions.get(i - 1), activations.get(i)).derivative()));
}
}
for (int i = 0; i < deltas.length; i++) {
deltaRet.add(new Pair<>(deltas[i], preCons[i]));
}
return deltaRet;
}
@Override
public MultiLayerNetwork clone() {
MultiLayerNetwork ret;
try {
Constructor<MultiLayerNetwork> constructor = (Constructor) getClass().getDeclaredConstructor(MultiLayerConfiguration.class);
ret = constructor.newInstance(getLayerWiseConfigurations().clone());
ret.update(this);
ret.setParameters(params().dup());
} catch (Exception e) {
throw new IllegalStateException("Unable to clone network",e);
}
return ret;
}
/**
* Returns a 1 x m vector where the vector is composed of
* a flattened vector of all of the weights for the
* various neuralNets(w,hbias NOT VBIAS) and output layer
*
* @return the params for this neural net
*/
public INDArray params(boolean backwardOnly) {
if(backwardOnly) return params();
// if(params != null)
// return params;
List<INDArray> params = new ArrayList<>();
for (Layer layer: getLayers()){
INDArray layerParams;
if( layer instanceof BasePretrainNetwork && backwardOnly)
layerParams = ((BasePretrainNetwork) layer).paramsBackprop();
else
layerParams = layer.params();
if(layerParams != null) params.add(layerParams); //may be null: subsampling etc layers
}
return Nd4j.toFlattened('f', params);
}
/**
* Returns a 1 x m vector where the vector is composed of
* a flattened vector of all of the weights for the
* various neuralNets(w,hbias NOT VBIAS) and output layer
*
* @return the params for this neural net
*/
@Override
public INDArray params() {
return flattenedParams;
}
/**
* Set the parameters for this model.
* This expects a linear ndarray
* which then be unpacked internally
* relative to the expected ordering of the model
*
* @param params the parameters for the model
*/
@Override
public void setParams(INDArray params) {
if(flattenedParams == params) return; //No op
if(flattenedParams != null && params.length() == flattenedParams.length()){
flattenedParams.assign(params);
} else {
if(flattenedParams == null) flattenedParams = params.dup();
int idx = 0;
for (int i = 0; i < getLayers().length; i++) {
Layer layer = getLayer(i);
int range = (layer instanceof BasePretrainNetwork ?
((BasePretrainNetwork<?>)layer).numParamsBackprop() : layer.numParams());
if(range <= 0) continue; //Some layers: no parameters (subsampling, etc)
INDArray get = params.get(NDArrayIndex.point(0),NDArrayIndex.interval(idx, range + idx));
layer.setParams(get);
idx += range;
}
}
}
@Override
public void setParamsViewArray(INDArray params) {
throw new UnsupportedOperationException("Not yet implemented");
}
@Override
public void setBackpropGradientsViewArray(INDArray gradients) {
throw new UnsupportedOperationException("Not yet implemented");
}
/**
* Returns a 1 x m vector where the vector is composed of
* a flattened vector of all of the weights for the
* various neuralNets and output layer
*
* @return the params for this neural net
*/
@Override
public int numParams() {
return numParams(false);
}
@Override
public int numParams(boolean backwards) {
int length = 0;
for (int i = 0; i < layers.length; i++)
length += layers[i].numParams(backwards);
return length;
}
/**
* Packs a set of matrices in to one vector,
* where the matrices in this case are the w,hbias at each layer
* and the output layer w,bias
*
* @return a singular matrix of all of the neuralNets packed in to one matrix
*/
public INDArray pack() {
return params();
}
/**
* Packs a set of matrices in to one vector
*
* @param layers the neuralNets to pack
* @return a singular matrix of all of the neuralNets packed in to one matrix
*/
public INDArray pack(List<Pair layers) {
List<INDArray> list = new ArrayList<>();
for (Pair<INDArray, INDArray> layer : layers) {
list.add(layer.getFirst());
list.add(layer.getSecond());
}
return Nd4j.toFlattened(list);
}
/**
* Sets the input and labels and returns a score for the prediction
* wrt true labels
*
* @param data the data to score
* @return the score for the given input,label pairs
*/
@Override
public double f1Score(org.nd4j.linalg.dataset.api.DataSet data) {
return f1Score(data.getFeatureMatrix(), data.getLabels());
}
/**
* Unpacks a parameter matrix in to a
* transform of pairs(w,hbias)
* triples with layer wise
*
* @param param the param vector
* @return a segmented list of the param vector
*/
public List<Pair unPack(INDArray param) {
//more sanity checks!
if (param.size(0) != 1)
param = param.reshape(1, param.length());
List<Pair ret = new ArrayList<>();
int curr = 0;
for (int i = 0; i < layers.length; i++) {
int layerLength = layers[i].getParam(DefaultParamInitializer.WEIGHT_KEY).length() + layers[i].getParam(DefaultParamInitializer.BIAS_KEY).length();
INDArray subMatrix = param.get(NDArrayIndex.interval(curr, curr + layerLength));
INDArray weightPortion = subMatrix.get(NDArrayIndex.interval(0, layers[i].getParam(DefaultParamInitializer.WEIGHT_KEY).length()));
int beginHBias = layers[i].getParam(DefaultParamInitializer.WEIGHT_KEY).length();
int endHbias = subMatrix.length();
INDArray hBiasPortion = subMatrix.get(NDArrayIndex.interval(beginHBias, endHbias));
int layerLengthSum = weightPortion.length() + hBiasPortion.length();
if (layerLengthSum != layerLength) {
if (hBiasPortion.length() != layers[i].getParam(DefaultParamInitializer.BIAS_KEY).length())
throw new IllegalStateException("Hidden bias on layer " + i + " was off");
if (weightPortion.length() != layers[i].getParam(DefaultParamInitializer.WEIGHT_KEY).length())
throw new IllegalStateException("Weight portion on layer " + i + " was off");
}
ret.add(new Pair<>(weightPortion.reshape(layers[i].getParam(DefaultParamInitializer.WEIGHT_KEY).size(0), layers[i].getParam(DefaultParamInitializer.WEIGHT_KEY).columns()), hBiasPortion.reshape(layers[i].getParam(DefaultParamInitializer.BIAS_KEY).size(0), layers[i].getParam(DefaultParamInitializer.BIAS_KEY).columns())));
curr += layerLength;
}
return ret;
}
@Override
public void fit(DataSetIterator iterator) {
DataSetIterator iter;
// we're wrapping all iterators into AsyncDataSetIterator to provide background prefetch
if (!(iterator instanceof AsyncDataSetIterator || iterator instanceof ListDataSetIterator || iterator instanceof MultipleEpochsIterator)) {
iter = new AsyncDataSetIterator(iterator, 10);
} else iter = iterator;
if (layerWiseConfigurations.isPretrain()) {
pretrain(iter);
iter.reset();
while (iter.hasNext()) {
DataSet next = iter.next();
if (next.getFeatureMatrix() == null || next.getLabels() == null)
break;
setInput(next.getFeatureMatrix());
setLabels(next.getLabels());
finetune();
}
}
if (layerWiseConfigurations.isBackprop()) {
if(layerWiseConfigurations.isPretrain())
iter.reset();
update(TaskUtils.buildTask(iter));
iter.reset();
while (iter.hasNext()) {
DataSet next = iter.next();
if (next.getFeatureMatrix() == null || next.getLabels() == null)
break;
boolean hasMaskArrays = next.hasMaskArrays();
if(layerWiseConfigurations.getBackpropType() == BackpropType.TruncatedBPTT) {
doTruncatedBPTT(next.getFeatureMatrix(),next.getLabels(),next.getFeaturesMaskArray(),next.getLabelsMaskArray());
}
else {
if(hasMaskArrays) setLayerMaskArrays(next.getFeaturesMaskArray(), next.getLabelsMaskArray());
setInput(next.getFeatureMatrix());
setLabels(next.getLabels());
if( solver == null ){
solver = new Solver.Builder()
.configure(conf())
.listeners(getListeners())
.model(this).build();
}
solver.optimize();
}
if(hasMaskArrays) clearLayerMaskArrays();
}
}
}
/** Calculate and set gradients for MultiLayerNetwork, based on OutputLayer and labels*/
protected void backprop() {
if(flattenedGradients == null) initGradientsView();
Pair<Gradient,INDArray> pair = calcBackpropGradients(null, true);
this.gradient = (pair == null ? null : pair.getFirst());
this.epsilon = (pair == null ? null : pair.getSecond());
}
/** Calculate gradients and errors. Used in two places:
* (a) backprop (for standard multi layer network learning)
* (b) backpropGradient (layer method, for when MultiLayerNetwork is used as a layer)
* @param epsilon Errors (technically errors .* activations). Not used if withOutputLayer = true
* @param withOutputLayer if true: assume last layer is output layer, and calculate errors based on labels. In this
* case, the epsilon input is not used (may/should be null).
* If false: calculate backprop gradients
* @return Gradients and the error (epsilon) at the input
*/
protected Pair<Gradient,INDArray> calcBackpropGradients(INDArray epsilon, boolean withOutputLayer) {
if(flattenedGradients == null) initGradientsView();
String multiGradientKey;
Gradient gradient = new DefaultGradient(flattenedGradients);
Layer currLayer;
//calculate and apply the backward gradient for every layer
/**
* Skip the output layer for the indexing and just loop backwards updating the coefficients for each layer.
* (when withOutputLayer == true)
*
* Activate applies the activation function for each layer and sets that as the input for the following layer.
*
* Typical literature contains most trivial case for the error calculation: wT * weights
* This interpretation transpose a few things to get mini batch because ND4J is rows vs columns organization for params
*/
int numLayers = getnLayers();
//Store gradients is a list; used to ensure iteration order in DefaultGradient linked hash map. i.e., layer 0 first instead of output layer
LinkedList<Triple gradientList = new LinkedList<>();
int layerFrom;
Pair<Gradient,INDArray> currPair;
if(withOutputLayer) {
if(!(getOutputLayer() instanceof BaseOutputLayer)) {
log.warn("Warning: final layer isn't output layer. You cannot use backprop without an output layer.");
return null;
}
BaseOutputLayer<?> outputLayer = (BaseOutputLayer) getOutputLayer();
if (labels == null)
throw new IllegalStateException("No labels found");
outputLayer.setLabels(labels);
currPair = outputLayer.backpropGradient(null);
for( Map.Entry<String, INDArray> entry : currPair.getFirst().gradientForVariable().entrySet()) {
String origName = entry.getKey();
multiGradientKey = String.valueOf(numLayers - 1) + "_" + origName;
gradientList.addLast(new Triple<>(multiGradientKey,entry.getValue(),currPair.getFirst().flatteningOrderForVariable(origName)));
}
if(getLayerWiseConfigurations().getInputPreProcess(numLayers-1) != null)
currPair = new Pair<> (currPair.getFirst(), this.layerWiseConfigurations.getInputPreProcess(numLayers - 1).backprop(currPair.getSecond(),getInputMiniBatchSize()));
layerFrom = numLayers-2;
} else {
currPair = new Pair<>(null,epsilon);
layerFrom = numLayers-1;
}
// Calculate gradients for previous layers & drops output layer in count
for(int j = layerFrom; j >= 0; j--) {
currLayer = getLayer(j);
currPair = currLayer.backpropGradient(currPair.getSecond());
LinkedList<Triple tempList = new LinkedList<>();
for(Map.Entry<String, INDArray> entry : currPair.getFirst().gradientForVariable().entrySet()) {
String origName = entry.getKey();
multiGradientKey = String.valueOf(j) + "_" + origName;
tempList.addFirst(new Triple<>(multiGradientKey,entry.getValue(), currPair.getFirst().flatteningOrderForVariable(origName)));
}
for(Triple<String,INDArray,Character> triple : tempList) gradientList.addFirst(triple);
//Pass epsilon through input processor before passing to next layer (if applicable)
if(getLayerWiseConfigurations().getInputPreProcess(j) != null)
currPair = new Pair<> (currPair.getFirst(), getLayerWiseConfigurations().getInputPreProcess(j).backprop(currPair.getSecond(),getInputMiniBatchSize()));
}
//Add gradients to Gradients (map), in correct order
for( Triple<String,INDArray,Character> triple : gradientList) {
gradient.setGradientFor(triple.getFirst(), triple.getSecond(), triple.getThird());
}
return new Pair<>(gradient,currPair.getSecond());
}
protected void doTruncatedBPTT(INDArray input, INDArray labels, INDArray featuresMaskArray, INDArray labelsMaskArray) {
if( input.rank() != 3 || labels.rank() != 3 ){
log.warn("Cannot do truncated BPTT with non-3d inputs or labels. Expect input with shape [miniBatchSize,nIn,timeSeriesLength], got "
+ Arrays.toString(input.shape()) + "\t" + Arrays.toString(labels.shape()));
return;
}
if( input.size(2) != labels.size(2) ){
log.warn("Input and label time series have different lengths: {} input length, {} label length", input.size(2), labels.size(2));
return;
}
int fwdLen = layerWiseConfigurations.getTbpttFwdLength();
update(TaskUtils.buildTask(input, labels));
int timeSeriesLength = input.size(2);
int nSubsets = timeSeriesLength / fwdLen;
if(fwdLen > timeSeriesLength) {
log.warn("Cannot do TBPTT: Truncated BPTT forward length (" + fwdLen + ") > input time series length (" + timeSeriesLength + ")");
return;
}
rnnClearPreviousState();
for( int i=0; i<nSubsets; i++ ){
int startTimeIdx = i*fwdLen;
int endTimeIdx = startTimeIdx + fwdLen;
INDArray inputSubset = input.get(NDArrayIndex.all(),NDArrayIndex.all(),NDArrayIndex.interval(startTimeIdx, endTimeIdx));
INDArray labelSubset = labels.get(NDArrayIndex.all(),NDArrayIndex.all(),NDArrayIndex.interval(startTimeIdx, endTimeIdx));
setInput(inputSubset);
setLabels(labelSubset);
INDArray featuresMaskSubset = null;
INDArray labelsMaskSubset = null;
if(featuresMaskArray != null){
featuresMaskSubset = featuresMaskArray.get(NDArrayIndex.all(), NDArrayIndex.interval(startTimeIdx,endTimeIdx));
}
if(labelsMaskArray != null){
labelsMaskSubset = labelsMaskArray.get(NDArrayIndex.all(), NDArrayIndex.interval(startTimeIdx,endTimeIdx));
}
if(featuresMaskSubset != null || labelsMaskSubset != null) setLayerMaskArrays(featuresMaskSubset,labelsMaskSubset);
if(solver == null) {
solver = new Solver.Builder()
.configure(conf())
.listeners(getListeners())
.model(this).build();
}
solver.optimize();
//Finally, update the state of the RNN layers:
updateRnnStateWithTBPTTState();
}
rnnClearPreviousState();
if(featuresMaskArray != null || labelsMaskArray != null) clearLayerMaskArrays();
}
public void updateRnnStateWithTBPTTState() {
for(int i=0; i<layers.length; i++){
if(layers[i] instanceof BaseRecurrentLayer) {
BaseRecurrentLayer<?> l = ((BaseRecurrentLayer)layers[i]);
l.rnnSetPreviousState(l.rnnGetTBPTTState());
}
else if(layers[i] instanceof MultiLayerNetwork) {
((MultiLayerNetwork)layers[i]).updateRnnStateWithTBPTTState();
}
}
}
/** Equivalent to backprop(), but calculates gradient for truncated BPTT instead. */
protected void truncatedBPTTGradient(){
if(flattenedGradients == null) initGradientsView();
String multiGradientKey;
gradient = new DefaultGradient();
Layer currLayer;
if(!(getOutputLayer() instanceof BaseOutputLayer)) {
log.warn("Warning: final layer isn't output layer. You cannot use backprop (truncated BPTT) without an output layer.");
return;
}
BaseOutputLayer<?> outputLayer = (BaseOutputLayer) getOutputLayer();
if(labels == null)
throw new IllegalStateException("No labels found");
if(outputLayer.conf().getLayer().getWeightInit() == WeightInit.ZERO){
throw new IllegalStateException("Output layer weights cannot be initialized to zero when using backprop.");
}
outputLayer.setLabels(labels);
//calculate and apply the backward gradient for every layer
int numLayers = getnLayers();
//Store gradients is a list; used to ensure iteration order in DefaultGradient linked hash map. i.e., layer 0 first instead of output layer
LinkedList<Pair gradientList = new LinkedList<>();
Pair<Gradient,INDArray> currPair = outputLayer.backpropGradient(null);
for( Map.Entry<String, INDArray> entry : currPair.getFirst().gradientForVariable().entrySet()) {
multiGradientKey = String.valueOf(numLayers - 1) + "_" + entry.getKey();
gradientList.addLast(new Pair<>(multiGradientKey,entry.getValue()));
}
if(getLayerWiseConfigurations().getInputPreProcess(numLayers-1) != null)
currPair = new Pair<> (currPair.getFirst(), this.layerWiseConfigurations.getInputPreProcess(numLayers - 1).backprop(currPair.getSecond(),getInputMiniBatchSize()));
// Calculate gradients for previous layers & drops output layer in count
for(int j = numLayers - 2; j >= 0; j--) {
currLayer = getLayer(j);
if(currLayer instanceof BaseRecurrentLayer){
currPair = ((BaseRecurrentLayer<?>)currLayer).tbpttBackpropGradient(currPair.getSecond(),layerWiseConfigurations.getTbpttBackLength());
} else {
currPair = currLayer.backpropGradient(currPair.getSecond());
}
LinkedList<Pair tempList = new LinkedList<>();
for(Map.Entry<String, INDArray> entry : currPair.getFirst().gradientForVariable().entrySet()) {
multiGradientKey = String.valueOf(j) + "_" + entry.getKey();
tempList.addFirst(new Pair<>(multiGradientKey,entry.getValue()));
}
for(Pair<String,INDArray> pair : tempList)
gradientList.addFirst(pair);
//Pass epsilon through input processor before passing to next layer (if applicable)
if(getLayerWiseConfigurations().getInputPreProcess(j) != null)
currPair = new Pair<> (currPair.getFirst(), getLayerWiseConfigurations().getInputPreProcess(j).backprop(currPair.getSecond(),getInputMiniBatchSize()));
}
//Add gradients to Gradients, in correct order
for( Pair<String,INDArray> pair : gradientList)
gradient.setGradientFor(pair.getFirst(), pair.getSecond());
}
/**
*
* @return
*/
public Collection<IterationListener> getListeners() {
return listeners;
}
@Override
public void setListeners(Collection<IterationListener> listeners) {
this.listeners = listeners;
if (layers == null) {
init();
}
for (Layer layer : layers) {
layer.setListeners(listeners);
}
if(solver != null){
solver.setListeners(listeners);
}
}
@Override
public void setListeners(IterationListener... listeners) {
Collection<IterationListener> cListeners = new ArrayList<>();
Collections.addAll(cListeners, listeners);
setListeners(cListeners);
}
/**
* Run SGD based on the given labels
*
*/
public void finetune() {
if(flattenedGradients == null) initGradientsView();
if (!(getOutputLayer() instanceof BaseOutputLayer)) {
log.warn("Output layer not instance of output layer returning.");
return;
}
if(labels == null)
throw new IllegalStateException("No labels found");
log.info("Finetune phase");
BaseOutputLayer output = (BaseOutputLayer) getOutputLayer();
if (output.conf().getOptimizationAlgo() != OptimizationAlgorithm.HESSIAN_FREE) {
feedForward();
output.fit(output.input(), labels);
} else {
throw new UnsupportedOperationException();
}
}
/**
* Returns the predictions for each example in the dataset
*
* @param d the matrix to predict
* @return the prediction for the dataset
*/
@Override
public int[] predict(INDArray d) {
INDArray output = output(d, Layer.TrainingMode.TEST);
int[] ret = new int[d.size(0)];
if (d.isRowVector()) ret[0] = Nd4j.getBlasWrapper().iamax(output);
else {
for (int i = 0; i < ret.length; i++)
ret[i] = Nd4j.getBlasWrapper().iamax(output.getRow(i));
}
return ret;
}
/**
* Return predicted label names
*
* @param dataSet to predict
* @return the predicted labels for the dataSet
*/
@Override
public List<String> predict(org.nd4j.linalg.dataset.api.DataSet dataSet) {
int[] intRet = predict(dataSet.getFeatureMatrix());
List<String> ret = new ArrayList<>();
for(int i=0; i < intRet.length; i++) {
ret.add(i,dataSet.getLabelName(intRet[i]));
}
return ret;
}
/**
* Returns the probabilities for each label
* for each example row wise
*
* @param examples the examples to classify (one example in each row)
* @return the likelihoods of each example and each label
*/
@Override
public INDArray labelProbabilities(INDArray examples) {
List<INDArray> feed = feedForward(examples);
BaseOutputLayer o = (BaseOutputLayer) getOutputLayer();
return o.labelProbabilities(feed.get(feed.size() - 1));
}
/**
* Fit the model
*
* @param data the examples to classify (one example in each row)
* @param labels the example labels(a binary outcome matrix)
*/
@Override
public void fit(INDArray data, INDArray labels) {
setInput(data);
setLabels(labels);
update(TaskUtils.buildTask(data, labels));
if (layerWiseConfigurations.isPretrain()) {
pretrain(data);
finetune();
}
if(layerWiseConfigurations.isBackprop()) {
if(layerWiseConfigurations.getBackpropType() == BackpropType.TruncatedBPTT) {
doTruncatedBPTT(data,labels,null,null);
}
else {
if( solver == null) {
solver = new Solver.Builder()
.configure(conf())
.listeners(getListeners())
.model(this).build();
}
solver.optimize();
}
}
}
/**
* Fit the unsupervised model
*
* @param data the examples to classify (one example in each row)
*/
@Override
public void fit(INDArray data) {
setInput(data);
update(TaskUtils.buildTask(data));
pretrain(data);
}
@Override
public void iterate(INDArray input) {
pretrain(input);
}
/**
* Fit the model
*
* @param data the data to train on
*/
@Override
public void fit(org.nd4j.linalg.dataset.api.DataSet data) {
if(layerWiseConfigurations.getBackpropType() == BackpropType.TruncatedBPTT) {
doTruncatedBPTT(data.getFeatureMatrix(),data.getLabels(),data.getFeaturesMaskArray(),data.getLabelsMaskArray());
} else {
//Standard training
boolean hasMaskArrays = data.hasMaskArrays();
if(hasMaskArrays) setLayerMaskArrays(data.getFeaturesMaskArray(), data.getLabelsMaskArray());
fit(data.getFeatureMatrix(), data.getLabels());
if(hasMaskArrays) clearLayerMaskArrays();
}
}
/**
* Fit the model
*
* @param examples the examples to classify (one example in each row)
* @param labels the labels for each example (the number of labels must match
*/
@Override
public void fit(INDArray examples, int[] labels) {
org.deeplearning4j.nn.conf.layers.OutputLayer layerConf =
(org.deeplearning4j.nn.conf.layers.OutputLayer) getOutputLayer().conf().getLayer();
fit(examples, FeatureUtil.toOutcomeMatrix(labels, layerConf.getNOut()));
}
/**
* Label the probabilities of the input
*
* @param input the input to label
* @param train whether the output
* is test or train. This mainly
* affect hyper parameters such as
* drop out where certain things should
* be applied with activations
* @return a vector of probabilities
* given each label.
* <p>
* This is typically of the form:
* [0.5, 0.5] or some other probability distribution summing to one
*/
public INDArray output(INDArray input, TrainingMode train) {
return output(input,train == TrainingMode.TRAIN);
}
/**
* Label the probabilities of the input
*
* @param input the input to label
* @param train whether the output
* is test or train. This mainly
* affect hyper parameters such as
* drop out where certain things should
* be applied with activations
* @return a vector of probabilities
* given each label.
* <p>
* This is typically of the form:
* [0.5, 0.5] or some other probability distribution summing to one
*/
public INDArray output(INDArray input, boolean train) {
List<INDArray> activations = feedForward(input, train);
//last activation is output
return activations.get(activations.size() - 1);
}
/** Calculate the output of the network, with masking arrays. The masking arrays are used in situations such
* as one-to-many and many-to-one recurrent neural network (RNN) designs, as well as for supporting time series
* of varying lengths within the same minibatch.
*/
public INDArray output(INDArray input, boolean train, INDArray featuresMask, INDArray labelsMask){
setLayerMaskArrays(featuresMask,labelsMask);
INDArray out = output(input, train);
clearLayerMaskArrays();
return out;
}
/**
* Label the probabilities of the input
*
* @param input the input to label
* @return a vector of probabilities
* given each label.
* <p>
* This is typically of the form:
* [0.5, 0.5] or some other probability distribution summing to one
*/
public INDArray output(INDArray input) {
return output(input, TrainingMode.TRAIN);
}
/**
* Label the probabilities of the input
*
* @param iterator test data to evaluate
* @return a vector of probabilities
* given each label.
* <p>
* This is typically of the form:
* [0.5, 0.5] or some other probability distribution summing to one
*/
public INDArray output(DataSetIterator iterator, boolean train) {
List<INDArray> outList = new ArrayList<>();
while(iterator.hasNext()){
DataSet next = iterator.next();
if (next.getFeatureMatrix() == null || next.getLabels() == null)
break;
INDArray features = next.getFeatures();
if(next.hasMaskArrays()){
INDArray fMask = next.getFeaturesMaskArray();
INDArray lMask = next.getLabelsMaskArray();
outList.add(this.output(features,train,fMask,lMask));
} else {
outList.add(output(features,train));
}
}
return Nd4j.vstack(outList.toArray(new INDArray[0]));
}
public INDArray output(DataSetIterator iterator) {
return output(iterator, false);
}
/**
* Reconstructs the input.
* This is equivalent functionality to a
* deep autoencoder.
*
* @param x the input to transform
* @param layerNum the layer to output for encoding
* @return a reconstructed matrix
* relative to the size of the last hidden layer.
* This is great for data compression and visualizing
* high dimensional data (or just doing dimensionality reduction).
* <p>
* This is typically of the form:
* [0.5, 0.5] or some other probability distribution summing to one
*/
public INDArray reconstruct(INDArray x, int layerNum) {
List<INDArray> forward = feedForward(x);
return forward.get(layerNum - 1);
}
/**
* Prints the configuration
*/
public void printConfiguration() {
StringBuilder sb = new StringBuilder();
int count = 0;
for (NeuralNetConfiguration conf : getLayerWiseConfigurations().getConfs()) {
sb.append(" Layer " + count++ + " conf " + conf);
}
log.info(sb.toString());
}
/**
* Assigns the parameters of this model to the ones specified by this
* network. This is used in loading from input streams, factory methods, etc
*
* @param network the network to getFromOrigin parameters from
*/
public void update(MultiLayerNetwork network) {
this.defaultConfiguration = (network.defaultConfiguration != null ? network.defaultConfiguration.clone() : null);
if(network.input != null) setInput(network.input.dup()); //Dup in case of dropout etc
this.labels = network.labels;
if(network.layers != null){
layers = new Layer[network.layers.length];
for( int i=0; i<layers.length; i++ ){
layers[i] = network.layers[i].clone();
}
} else {
this.layers = null;
}
if(network.solver != null){
//Network updater state: should be cloned over also
this.setUpdater(network.getUpdater().clone());
} else {
this.solver = null;
}
}
/**
* Sets the input and labels and returns a score for the prediction
* wrt true labels
*
* @param input the input to score
* @param labels the true labels
* @return the score for the given input,label pairs
*/
@Override
public double f1Score(INDArray input, INDArray labels) {
feedForward(input);
setLabels(labels);
Evaluation eval = new Evaluation();
eval.eval(labels, labelProbabilities(input));
return eval.f1();
}
/**
* Returns the number of possible labels
*
* @return the number of possible labels for this classifier
*/
@Override
public int numLabels() {
return labels.columns();
}
/**Sets the input and labels and returns a score for the prediction with respect to the true labels<br>
* This is equivalent to {@link #score(DataSet, boolean)} with training==true.
* @param data the data to score
* @return the score for the given input,label pairs
* @see #score(DataSet, boolean)
*/
public double score(DataSet data) {
return score(data,false);
}
/**Calculate the score (loss function) of the prediction with respect to the true labels<br>
* @param data data to calculate score for
* @param training If true: score during training. If false: score at test time. This can affect the application of
* certain features, such as dropout and dropconnect (which are applied at training time only)
* @return the score (value of the loss function)
*/
public double score(DataSet data, boolean training){
boolean hasMaskArray = data.hasMaskArrays();
if(hasMaskArray) setLayerMaskArrays(data.getFeaturesMaskArray(),data.getLabelsMaskArray());
// activation for output layer is calculated in computeScore
List<INDArray> activations = feedForwardToLayer(layers.length - 2, data.getFeatureMatrix(),training);
int n = activations.size();
setLabels(data.getLabels());
if( getOutputLayer() instanceof BaseOutputLayer ){
BaseOutputLayer<?> ol = (BaseOutputLayer)getOutputLayer();
INDArray olInput = activations.get(n-1);
if(getLayerWiseConfigurations().getInputPreProcess(n-1) != null){
olInput = getLayerWiseConfigurations().getInputPreProcess(n-1).preProcess(olInput,input.size(0));
}
ol.setInput(olInput); //Feedforward doesn't include output layer for efficiency
ol.setLabels(data.getLabels());
ol.computeScore(calcL1(),calcL2(), training);
this.score = ol.score();
} else {
log.warn("Cannot calculate score wrt labels without an OutputLayer");
return 0.0;
}
if(hasMaskArray) clearLayerMaskArrays();
return score();
}
public INDArray scoreExamples(DataSetIterator iter, boolean addRegularizationTerms){
List<INDArray> out = new ArrayList<>();
while(iter.hasNext()){
out.add(scoreExamples(iter.next(), addRegularizationTerms));
}
return Nd4j.toFlattened('f',out);
}
/**Calculate the score for each example in a DataSet individually. Unlike {@link #score(DataSet)} and {@link #score(DataSet, boolean)}
* this method does not average/sum over examples. This method allows for examples to be scored individually (at test time only), which
* may be useful for example for autoencoder architectures and the like.<br>
* Each row of the output (assuming addRegularizationTerms == true) is equivalent to calling score(DataSet) with a single example.
* @param data The data to score
* @param addRegularizationTerms If true: add l1/l2 regularization terms (if any) to the score. If false: don't add regularization terms
* @return An INDArray (column vector) of size input.numRows(); the ith entry is the score (loss value) of the ith example
*/
public INDArray scoreExamples(DataSet data, boolean addRegularizationTerms){
boolean hasMaskArray = data.hasMaskArrays();
if(hasMaskArray) setLayerMaskArrays(data.getFeaturesMaskArray(),data.getLabelsMaskArray());
feedForward(data.getFeatureMatrix(),false);
setLabels(data.getLabels());
INDArray out;
if( getOutputLayer() instanceof BaseOutputLayer ){
BaseOutputLayer<?> ol = (BaseOutputLayer)getOutputLayer();
ol.setLabels(data.getLabels());
double l1 = (addRegularizationTerms ? calcL1() : 0.0);
double l2 = (addRegularizationTerms ? calcL2() : 0.0);
out = ol.computeScoreForExamples(l1,l2);
} else {
throw new UnsupportedOperationException("Cannot calculate score wrt labels without an OutputLayer");
}
if(hasMaskArray) clearLayerMaskArrays();
return out;
}
@Override
public void fit() {
fit(input, labels);
}
@Override
public void update(INDArray gradient, String paramType) {
}
/**
* Score of the model (relative to the objective function)
*
* @return the score of the model (relative to the objective function)
*/
@Override
public double score() {
return score;
}
public void setScore(double score) {
this.score = score;
}
@Override
public void computeGradientAndScore() {
//Calculate activations (which are stored in each layer, and used in backprop)
if(layerWiseConfigurations.getBackpropType() == BackpropType.TruncatedBPTT) {
rnnActivateUsingStoredState(getInput(), true, true);
truncatedBPTTGradient();
}
else {
//First: do a feed-forward through the network
//Note that we don't actually need to do the full forward pass through the output layer right now; but we do
// need the input to the output layer to be set (such that backprop can be done)
List<INDArray> activations = feedForwardToLayer(layers.length-2,true);
INDArray actSecondLastLayer = activations.get(activations.size()-1);
if(layerWiseConfigurations.getInputPreProcess(layers.length-1) != null)
actSecondLastLayer = layerWiseConfigurations.getInputPreProcess(layers.length-1).preProcess(actSecondLastLayer,getInputMiniBatchSize());
getOutputLayer().setInput(actSecondLastLayer);
//Then: compute gradients
backprop();
}
score = ((BaseOutputLayer<?>)getOutputLayer()).computeScore(calcL1(),calcL2(), true);
}
@Override
public void accumulateScore(double accum) {
}
/**
* Clear the inputs. Clears optimizer state.
*/
public void clear() {
for (Layer layer : layers)
layer.clear();
input = null;
labels = null;
solver = null;
}
/**
* Averages the given logistic regression
* from a mini batch in to this one
*
* @param layer the logistic regression to average in to this one
* @param batchSize the batch size
*/
@Override
public void merge(Layer layer, int batchSize) {
throw new UnsupportedOperationException();
}
/**
* Merges this network with the other one.
* This is a weight averaging with the update of:
* a += b - a / n
* where a is a matrix on the network
* b is the incoming matrix and n
* is the batch size.
* This update is performed across the network neuralNets
* as well as hidden neuralNets and logistic neuralNets
*
* @param network the network to merge with
* @param batchSize the batch size (number of training examples)
* to average by
*/
public void merge(MultiLayerNetwork network, int batchSize) {
if (network.layers.length != layers.length)
throw new IllegalArgumentException("Unable to merge networks that are not of equal length");
for (int i = 0; i < getnLayers(); i++) {
Layer n = layers[i];
Layer otherNetwork = network.layers[i];
n.merge(otherNetwork, batchSize);
}
getOutputLayer().merge(network.getOutputLayer(), batchSize);
}
/**
* Note that if input isn't null
* and the neuralNets are null, this is a way
* of initializing the neural network
*
* @param input
*/
public void setInput(INDArray input) {
this.input = input;
if (this.layers == null)
this.initializeLayers(getInput());
if(input != null) {
if(input.length() == 0) throw new IllegalArgumentException("Invalid input: length 0 (shape: " + Arrays.toString(input.shape()) +")");
setInputMiniBatchSize(input.size(0));
}
}
private void initMask() {
setMask(Nd4j.ones(1, pack().length()));
}
/**
* Get the output layer
*
* @return
*/
public Layer getOutputLayer() {
return getLayers()[getLayers().length - 1];
}
/**
* Sets parameters for the model.
* This is used to manipulate the weights and biases across
* all neuralNets (including the output layer)
*
* @param params a parameter vector equal 1,numParameters
*/
public void setParameters(INDArray params) {
setParams(params);
}
public void applyLearningRateScoreDecay() {
for (Layer layer: layers) {
if (!layer.conf().getLearningRateByParam().isEmpty()) {
for (Map.Entry<String, Double> lrPair : layer.conf().getLearningRateByParam().entrySet()) {
layer.conf().setLearningRateByParam(lrPair.getKey(),
lrPair.getValue() * (layer.conf().getLrPolicyDecayRate() + Nd4j.EPS_THRESHOLD));
}
}
}
}
/**
* Feed forward with the r operator
*
* @param v the v for the r operator
* @return the activations based on the r operator
*/
public List<INDArray> feedForwardR(List acts, INDArray v) {
List<INDArray> R = new ArrayList<>();
R.add(Nd4j.zeros(input.size(0), input.columns()));
List<Pair vWvB = unPack(v);
List<INDArray> W = MultiLayerUtil.weightMatrices(this);
for (int i = 0; i < layers.length; i++) {
String derivative = getLayers()[i].conf().getLayer().getActivationFunction();
//R[i] * W[i] + acts[i] * (vW[i] + vB[i]) .* f'([acts[i + 1])
R.add(R.get(i).mmul(W.get(i)).addi(acts.get(i)
.mmul(vWvB.get(i).getFirst().addiRowVector(vWvB.get(i).getSecond())))
.muli((Nd4j.getExecutioner().execAndReturn(Nd4j.getOpFactory().createTransform(derivative, acts.get(i + 1)).derivative()))));
}
return R;
}
public NeuralNetConfiguration getDefaultConfiguration() {
return defaultConfiguration;
}
public INDArray getLabels() {
return labels;
}
public INDArray getInput() {
return input;
}
/**
*
* @param labels
*/
public void setLabels(INDArray labels) {
this.labels = labels;
}
/**
* Get the number of layers in the network
*
* @return the number of layers in the network
*/
public int getnLayers() {
return layerWiseConfigurations.getConfs().size();
}
/**
*
* @return
*/
public Layer[] getLayers() {
return layers;
}
public Layer getLayer(int i) {
return layers[i];
}
public Layer getLayer(String name){
return layerMap.get(name);
}
public List<String> getLayerNames(){
return new ArrayList<>(layerMap.keySet());
}
public void setLayers(Layer[] layers) {
this.layers = layers;
}
public INDArray getMask() {
return mask;
}
public void setMask(INDArray mask) {
this.mask = mask;
}
//==========
//Layer methods
@Override
public Gradient error(INDArray errorSignal) {
throw new UnsupportedOperationException();
}
@Override
public Type type() {
return Type.MULTILAYER;
}
@Override
public INDArray derivativeActivation(INDArray input) {
throw new UnsupportedOperationException();
}
@Override
public Gradient calcGradient(Gradient layerError, INDArray activation) {
throw new UnsupportedOperationException();
}
@Override
public INDArray preOutput(INDArray x) {
INDArray lastLayerActivation = x;
for( int i=0; i<layers.length-1; i++ ){
if(getLayerWiseConfigurations().getInputPreProcess(i) != null)
lastLayerActivation = getLayerWiseConfigurations().getInputPreProcess(i).preProcess(lastLayerActivation,getInputMiniBatchSize());
lastLayerActivation = layers[i].activate(lastLayerActivation);
}
if(getLayerWiseConfigurations().getInputPreProcess(layers.length-1) != null)
lastLayerActivation = getLayerWiseConfigurations().getInputPreProcess(layers.length-1).preProcess(lastLayerActivation,getInputMiniBatchSize());
return layers[layers.length-1].preOutput(lastLayerActivation);
}
@Override
public INDArray preOutput(INDArray x, TrainingMode training) {
return preOutput(x, training == TrainingMode.TRAIN);
}
@Override
public INDArray activate(TrainingMode training) {
return activate(training == TrainingMode.TRAIN);
}
@Override
public INDArray activate(INDArray input, TrainingMode training) {
return activate(input,training == TrainingMode.TRAIN);
}
@Override
public Layer transpose() {
throw new UnsupportedOperationException();
}
@Override
public Pair<Gradient,INDArray> backpropGradient(INDArray epsilon) {
if(layers[layers.length-1] instanceof BaseOutputLayer<?> )
throw new UnsupportedOperationException("Cannot calculate gradients based on epsilon with OutputLayer");
return calcBackpropGradients(epsilon, false);
}
@Override
public void setIndex(int index){
layerIndex = index;
}
@Override
public int getIndex(){
return layerIndex;
}
@Override
public double calcL2() {
double l2 = 0.0;
for( int i=0; i<layers.length; i++ ){
l2 += layers[i].calcL2();
}
return l2;
}
@Override
public double calcL1() {
double l1 = 0.0;
for( int i=0; i<layers.length; i++ ){
l1 += layers[i].calcL1();
}
return l1;
}
@Override
public void update(Gradient gradient) {
throw new UnsupportedOperationException();
}
@Override
public INDArray preOutput(INDArray x, boolean training) {
throw new UnsupportedOperationException();
}
@Override
public INDArray activate(boolean training) {
throw new UnsupportedOperationException();
}
@Override
public INDArray activate(INDArray input, boolean training) {
throw new UnsupportedOperationException();
}
@Override
public void setInputMiniBatchSize(int size){
if(layers != null)
for(Layer l : layers)
l.setInputMiniBatchSize(size);
}
@Override
public int getInputMiniBatchSize(){
return input.size(0);
}
@Override
public void setMaskArray(INDArray maskArray) {
throw new UnsupportedOperationException();
}
/**
*
* If this MultiLayerNetwork contains one or more RNN layers: conduct forward pass (prediction)
* but using previous stored state for any RNN layers. The activations for the final step are
* also stored in the RNN layers for use next time rnnTimeStep() is called.<br>
* This method can be used to generate output one or more steps at a time instead of always having to do
* forward pass from t=0. Example uses are for streaming data, and for generating samples from network output
* one step at a time (where samples are then fed back into the network as input)<br>
* If no previous state is present in RNN layers (i.e., initially or after calling rnnClearPreviousState()),
* the default initialization (usually 0) is used.<br>
* Supports mini-batch (i.e., multiple predictions/forward pass in parallel) as well as for single examples.<br>
* @param input Input to network. May be for one or multiple time steps. For single time step:
* input has shape [miniBatchSize,inputSize] or [miniBatchSize,inputSize,1]. miniBatchSize=1 for single example.<br>
* For multiple time steps: [miniBatchSize,inputSize,inputTimeSeriesLength]
* @return Output activations. If output is RNN layer (such as RnnOutputLayer): if input has shape [miniBatchSize,inputSize]
* i.e., is 2d, output has shape [miniBatchSize,outputSize] (i.e., also 2d).<br>
* Otherwise output is 3d [miniBatchSize,outputSize,inputTimeSeriesLength] when using RnnOutputLayer.
*/
public INDArray rnnTimeStep(INDArray input) {
this.setInputMiniBatchSize(input.size(0)); //Necessary for preprocessors/reshaping
boolean inputIs2d = input.rank()==2;
for( int i = 0; i < layers.length; i++) {
if(getLayerWiseConfigurations().getInputPreProcess(i) != null)
input = getLayerWiseConfigurations().getInputPreProcess(i).preProcess(input,getInputMiniBatchSize());
if(layers[i] instanceof BaseRecurrentLayer){
input = ((BaseRecurrentLayer<?>)layers[i]).rnnTimeStep(input);
} else if(layers[i] instanceof MultiLayerNetwork){
input = ((MultiLayerNetwork)layers[i]).rnnTimeStep(input);
} else {
input = layers[i].activate(input, false);
}
}
if(inputIs2d && input.rank()==3 && layers[layers.length-1].type() == Type.RECURRENT){
//Return 2d output with shape [miniBatchSize,nOut]
// instead of 3d output with shape [miniBatchSize,nOut,1]
return input.tensorAlongDimension(0,1,0);
}
return input;
}
/**Get the state of the RNN layer, as used in rnnTimeStep().
* @param layer Number/index of the layer.
* @return Hidden state, or null if layer is not an RNN layer
*/
public Map<String,INDArray> rnnGetPreviousState(int layer){
if(layer < 0 || layer >= layers.length ) throw new IllegalArgumentException("Invalid layer number");
if( !(layers[layer] instanceof BaseRecurrentLayer) ) throw new IllegalArgumentException("Layer is not an RNN layer");
return ((BaseRecurrentLayer<?>)layers[layer]).rnnGetPreviousState();
}
/**Set the state of the RNN layer.
* @param layer The number/index of the layer.
* @param state The state to set the specified layer to
*/
public void rnnSetPreviousState(int layer, Map<String,INDArray> state){
if(layer < 0 || layer >= layers.length ) throw new IllegalArgumentException("Invalid layer number");
if( !(layers[layer] instanceof BaseRecurrentLayer) ) throw new IllegalArgumentException("Layer is not an RNN layer");
BaseRecurrentLayer<?> r = (BaseRecurrentLayer)layers[layer];
r.rnnSetPreviousState(state);
}
/** Clear the previous state of the RNN layers (if any).
*/
public void rnnClearPreviousState(){
if( layers == null ) return;
for( int i=0; i<layers.length; i++ ){
if( layers[i] instanceof BaseRecurrentLayer ) ((BaseRecurrentLayer<?>)layers[i]).rnnClearPreviousState();
else if( layers[i] instanceof MultiLayerNetwork ){
((MultiLayerNetwork)layers[i]).rnnClearPreviousState();
}
}
}
/** Similar to rnnTimeStep and feedForward() methods. Difference here is that this method:<br>
* (a) like rnnTimeStep does forward pass using stored state for RNN layers, and<br>
* (b) unlike rnnTimeStep does not modify the RNN layer state<br>
* Therefore multiple calls to this method with the same input should have the same output.<br>
* Typically used during training only. Use rnnTimeStep for prediction/forward pass at test time.
* @param input Input to network
* @param training Whether training or not
* @param storeLastForTBPTT set to true if used as part of truncated BPTT training
* @return Activations for each layer (including input, as per feedforward() etc)
*/
public List<INDArray> rnnActivateUsingStoredState(INDArray input, boolean training, boolean storeLastForTBPTT) {
INDArray currInput = input;
List<INDArray> activations = new ArrayList<>();
activations.add(currInput);
for( int i=0; i<layers.length; i++ ){
if(getLayerWiseConfigurations().getInputPreProcess(i) != null)
currInput = getLayerWiseConfigurations().getInputPreProcess(i).preProcess(currInput,input.size(0));
if(layers[i] instanceof BaseRecurrentLayer){
currInput = ((BaseRecurrentLayer<?>)layers[i]).rnnActivateUsingStoredState(currInput,training,storeLastForTBPTT);
} else if(layers[i] instanceof MultiLayerNetwork){
List<INDArray> temp = ((MultiLayerNetwork)layers[i]).rnnActivateUsingStoredState(currInput, training, storeLastForTBPTT);
currInput = temp.get(temp.size()-1);
} else {
currInput = layers[i].activate(currInput, training);
}
activations.add(currInput);
}
return activations;
}
/** Get the updater for this MultiLayerNetwork
* @return Updater for MultiLayerNetwork
*/
public synchronized Updater getUpdater() {
if(solver == null){
solver = new Solver.Builder()
.configure(conf())
.listeners(getListeners())
.model(this).build();
solver.getOptimizer().setUpdater(UpdaterCreator.getUpdater(this));
}
return solver.getOptimizer().getUpdater();
}
/** Set the updater for the MultiLayerNetwork */
public void setUpdater(Updater updater) {
if(solver == null) {
solver = new Solver.Builder()
.configure(conf())
.listeners(getListeners())
.model(this).build();
}
solver.getOptimizer().setUpdater(updater);
}
/**Set the mask arrays for features and labels. Mask arrays are typically used in situations such as one-to-many
* and many-to-one learning with recurrent neural networks, as well as for supporting time series of varying lengths
* within the same minibatch.<br>
* For example, with RNN data sets with input of shape [miniBatchSize,nIn,timeSeriesLength] and outputs of shape
* [miniBatchSize,nOut,timeSeriesLength], the features and mask arrays will have shape [miniBatchSize,timeSeriesLength]
* and contain values 0 or 1 at each element (to specify whether a given input/example is present - or merely padding -
* at a given time step).<br>
* <b>NOTE: This method is not usually used directly. Instead, methods such as {@link #feedForward(INDArray, INDArray, INDArray)}
* and {@link #output(INDArray, boolean, INDArray, INDArray)} handle setting of masking internally.
* @param featuresMaskArray Mask array for features (input)
* @param labelsMaskArray Mask array for labels (output)
* @see #clearLayerMaskArrays()
*/
public void setLayerMaskArrays(INDArray featuresMaskArray, INDArray labelsMaskArray){
if(featuresMaskArray != null){
//feedforward layers below a RNN layer: need the input (features) mask array
//Reason: even if the time series input is zero padded, the output from the dense layers are
// non-zero (i.e., activationFunction(0*weights + bias) != 0 in general)
//This assumes that the time series input is masked - i.e., values are 0 at the padded time steps,
// so we don't need to do anything for the recurrent layer
//Now, if mask array is 2d -> need to reshape to 1d (column vector) in the exact same order
// as is done for 3d -> 2d time series reshaping
INDArray reshapedFeaturesMask = TimeSeriesUtils.reshapeTimeSeriesMaskToVector(featuresMaskArray);
for( int i=0; i<layers.length-1; i++ ){
Type t = layers[i].type();
if( t == Type.CONVOLUTIONAL || t == Type.FEED_FORWARD ){
layers[i].setMaskArray(reshapedFeaturesMask);
} else if( t == Type.RECURRENT ) break;
}
}
if(labelsMaskArray != null ){
if(!(layers[layers.length-1] instanceof BaseOutputLayer) ) return;
layers[layers.length-1].setMaskArray(labelsMaskArray);
}
}
/** Remove the mask arrays from all layers.<br>
* See {@link #setLayerMaskArrays(INDArray, INDArray)} for details on mask arrays.
*/
public void clearLayerMaskArrays(){
for (Layer layer : layers) {
layer.setMaskArray(null);
}
}
/**Evaluate the network (classification performance)
* @param iterator Iterator to evaluate on
* @return Evaluation object; results of evaluation on all examples in the data set
*/
public Evaluation evaluate(DataSetIterator iterator) {
return evaluate(iterator, null);
}
/** Evaluate the network on the provided data set. Used for evaluating the performance of classifiers
* @param iterator Data to undertake evaluation on
* @return Evaluation object, summarizing returs of the evaluation
*/
public Evaluation evaluate(DataSetIterator iterator, List<String> labelsList){
if(layers == null || !(layers[layers.length-1] instanceof BaseOutputLayer)){
throw new IllegalStateException("Cannot evaluate network with no output layer");
}
if (labelsList == null)
labelsList = iterator.getLabels();
Evaluation e = (labelsList == null)? new Evaluation(): new Evaluation(labelsList);
while(iterator.hasNext()){
DataSet next = iterator.next();
if (next.getFeatureMatrix() == null || next.getLabels() == null)
break;
INDArray features = next.getFeatures();
INDArray labels = next.getLabels();
INDArray out;
if(next.hasMaskArrays()){
INDArray fMask = next.getFeaturesMaskArray();
INDArray lMask = next.getLabelsMaskArray();
out = this.output(features,false,fMask,lMask);
//Assume this is time series data. Not much point having a mask array for non TS data
if(lMask != null){
e.evalTimeSeries(labels,out,lMask);
} else {
e.evalTimeSeries(labels,out);
}
} else {
out = this.output(features,false);
if(labels.rank() == 3 ) e.evalTimeSeries(labels,out);
else e.eval(labels,out);
}
}
return e;
}
private void update(Task task) {
if (!initDone) {
initDone = true;
Heartbeat heartbeat = Heartbeat.getInstance();
task = ModelSerializer.taskByModel(this);
Environment env = EnvironmentUtils.buildEnvironment();
heartbeat.reportEvent(Event.STANDALONE, env, task);
}
}
}
Other Java examples (source code examples)
Here is a short list of links related to this Java MultiLayerNetwork.java source code file:
|