[net] subdivisions=8 inputs=256 batch = 128 momentum=0.9 decay=0.001 max_batches = 2000 time_steps=576 learning_rate=0.5 policy=steps burn_in=10 steps=1000,1500 scales=.1,.1 [lstm] batch_normalize=1 output = 1024 [lstm] batch_normalize=1 output = 1024 [lstm] batch_normalize=1 output = 1024 [connected] output=256 activation=leaky [softmax] [cost] type=sse