Skip to content

Instantly share code, notes, and snippets.

hiddenWidth = 127;
MultiLayerConfiguration conf = new NeuralNetConfiguration.Builder()
.optimizationAlgo(OptimizationAlgorithm.STOCHASTIC_GRADIENT_DESCENT).iterations(1)
.learningRate(0.05476383804339002) // 0.02 with BPTT
.momentum(0.5358333571512998)
.seed(12345)
.regularization(true)
.l2(0.016384859214629816)
.dropOut(0.8145498131858961)
hiddenWidth = 127;
MultiLayerConfiguration conf = new NeuralNetConfiguration.Builder()
.optimizationAlgo(OptimizationAlgorithm.STOCHASTIC_GRADIENT_DESCENT).iterations(1)
.learningRate(0.05476383804339002)
.momentum(0.5358333571512998)
.seed(12345)
.regularization(true)
.l2(0.016384859214629816)
.dropOut(0.8145498131858961)
{
"backprop" : true,
"backpropType" : "TruncatedBPTT",
"confs" : [ {
"extraArgs" : [ ],
"l1ByParam" : {
"b" : 0.0,
"RW" : 0.0,
"W" : 0.0
},
#
# A fatal error has been detected by the Java Runtime Environment:
#
# SIGSEGV (0xb) at pc=0x00007f6575c6cd34, pid=15404, tid=0x00007f65f482c700
#
# JRE version: Java(TM) SE Runtime Environment (8.0_101-b13) (build 1.8.0_101-b13)
# Java VM: Java HotSpot(TM) 64-Bit Server VM (25.101-b13 mixed mode linux-amd64 compressed oops)
# Problematic frame:
# C [libnd4j.so+0x1bd34] _ZN9functions19pairwise_transforms17PairWiseTransformIdE4execIN7simdOps4CopyIdEEEEvPdxS7_xS7_xS7_x._omp_fn.54+0x104
#
#
# SIGSEGV (0xb) at pc=0x00007f6575c6cd34, pid=15404, tid=0x00007f65f482c700
#
# JRE version: Java(TM) SE Runtime Environment (8.0_101-b13) (build 1.8.0_101-b13)
# Java VM: Java HotSpot(TM) 64-Bit Server VM (25.101-b13 mixed mode linux-amd64 compressed oops)
# Problematic frame:
# C [libnd4j.so+0x1bd34] _ZN9functions19pairwise_transforms17PairWiseTransformIdE4execIN7simdOps4CopyIdEEEEvPdxS7_xS7_xS7_x._omp_fn.54+0x104
#
# Failed to write core dump. Core dumps have been disabled. To enable core dumping, try "ulimit -c unlimited" before starting Java again
#
package org.deeplearning4j.examples.arbiter;
import org.deeplearning4j.arbiter.DL4JConfiguration;
import org.deeplearning4j.arbiter.MultiLayerSpace;
import org.deeplearning4j.arbiter.data.DataSetIteratorProvider;
import org.deeplearning4j.arbiter.layers.DenseLayerSpace;
import org.deeplearning4j.arbiter.layers.OutputLayerSpace;
import org.deeplearning4j.arbiter.optimize.api.CandidateGenerator;
import org.deeplearning4j.arbiter.optimize.api.OptimizationResult;
import org.deeplearning4j.arbiter.optimize.api.ParameterSpace;
INFO [2016-12-12 19:11:54,220] org.deeplearning4j.arbiter.optimize.runner.listener.runner.LoggingOptimizationRunnerStatusListener: Optimization runner: Initialized.
WARN [2016-12-12 19:13:31,720] org.deeplearning4j.arbiter.optimize.runner.BaseOptimizationRunner: Task failed
! java.lang.OutOfMemoryError: Physical memory usage is too high (21339664384 > Pointer.maxPhysicalBytes)
! at org.bytedeco.javacpp.Pointer.deallocator(Pointer.java:547)
! at org.bytedeco.javacpp.Pointer.init(Pointer.java:121)
! at org.bytedeco.javacpp.PointerPointer.allocateArray(Native Method)
! at org.bytedeco.javacpp.PointerPointer.<init>(PointerPointer.java:118)
! ... 21 common frames omitted
! Causing: java.lang.OutOfMemoryError: Cannot allocate new PointerPointer(4), totalBytes = 167141998, physicalBytes = 21169049600
! at org.bytedeco.javacpp.PointerPointer.<init>(PointerPointer.java:126)
// load net, normalizer....
net.rnnClearPreviousState();
// init Network with train history first
for (List<Integer> trainFeatures : trainFeatureValues) {
// construct features INDArray
int numExamples = 1;
int numFeatures = values.size();
int tsLength = 1;
<project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd">
<modelVersion>4.0.0</modelVersion>
<groupId>at.knowcenter.porsche</groupId>
<artifactId>dl4j-lib</artifactId>
<version>1.0-SNAPSHOT</version>
<name>Porsche DL Lib</name>
<description>Forecasting lib using dl4j</description>
public class RNNServiceImpl {
public CandidateGenerator<DL4JConfiguration> getCandidateGenerator() {
ParameterSpace<Double> learningRateHyperparam = new ContinuousParameterSpace(0.0001, 0.1);
ParameterSpace<Double> momentumHyperparam = new ContinuousParameterSpace(0.1, 0.9);
ParameterSpace<Double> dropoutHyperparam = new ContinuousParameterSpace(0.1, 0.9);
ParameterSpace<Integer> layerSizeHyperparam = new IntegerParameterSpace(16,256);
ParameterSpace<Integer> tbpttLengthHyperparam = new IntegerParameterSpace(1,100);
// layerspace builder for RNNs does not function as expected with chaining method calls?