From 17bfe22dc7d11ed496f233093ee99fff7c04f590 Mon Sep 17 00:00:00 2001 From: kaisheny Date: Fri, 29 May 2015 22:45:24 -0700 Subject: [PATCH] Experiment setup update for G2P --- ExampleSetups/G2P/setups/global.lstm.config | 4 ++-- ...mpd.rnd.hiddenstate.2nets.500.100mb.2layers.config | 2 +- ExampleSetups/SLU/README.txt | 11 ++++++++++- ExampleSetups/SLU/rnnlu.config | 6 ++++-- 4 files changed, 17 insertions(+), 6 deletions(-) diff --git a/ExampleSetups/G2P/setups/global.lstm.config b/ExampleSetups/G2P/setups/global.lstm.config index 73890d313..b57f05948 100644 --- a/ExampleSetups/G2P/setups/global.lstm.config +++ b/ExampleSetups/G2P/setups/global.lstm.config @@ -1,7 +1,7 @@ #WorkDir=//speechstore5/transient/kaishengy/data/lts/Data/CNTK WorkDir=d:/exp/lts -#DataDir=d:/data/lts -DataDir=d:/data/ltsdbg +DataDir=d:/data/lts +#DataDir=d:/data/ltsdbg NdlDir=c:/dev/cntk5/ExampleSetups/G2P/setups PredictionModelFeatureDir=\\speechstore5\transient\kaishengy\exp\lts\result\expbilstmce300n\s4 ExpDir=\\speechstore5\transient\kaishengy\exp\lts\result\explstm diff --git a/ExampleSetups/G2P/setups/s2s.mpd.rnd.hiddenstate.2nets.500.100mb.2layers.config b/ExampleSetups/G2P/setups/s2s.mpd.rnd.hiddenstate.2nets.500.100mb.2layers.config index be0b6cf7f..85a46a10d 100644 --- a/ExampleSetups/G2P/setups/s2s.mpd.rnd.hiddenstate.2nets.500.100mb.2layers.config +++ b/ExampleSetups/G2P/setups/s2s.mpd.rnd.hiddenstate.2nets.500.100mb.2layers.config @@ -3,7 +3,7 @@ stderr=$LogDir$\ATIS\log command=LSTM -type=double +type=float LSTM=[ # this is the maximum size for the minibatch, since sequence minibatches are really just a single sequence diff --git a/ExampleSetups/SLU/README.txt b/ExampleSetups/SLU/README.txt index b8bbb1b35..1edceaaea 100644 --- a/ExampleSetups/SLU/README.txt +++ b/ExampleSetups/SLU/README.txt @@ -16,7 +16,7 @@ Run the command line with both globals.config and the desired config, separated * note that full paths to config files need to be provided if you are not inside the config directory * for example -* C:\dev\cntk5\CNTKSolution\x64\Release\cntk configFile=C:\dev\cntk5\ExampleSetups\SLU\globals.config+C:\dev\cntk5\ExampleSetups\SLU\rnnlu.config +* C:\dev\cntk5\x64\release\CNTK.exe configFile=C:\dev\cntk5\ExampleSetups\SLU\globals.config+C:\dev\cntk5\ExampleSetups\SLU\rnnlu.config Scoring * ./score.sh @@ -27,6 +27,15 @@ Path Definitions: ================= * globals.config [defines paths to feature and label files and experiments] +Check training loss +========================== +$ grep Finish log_LSTM_LSTMTest.log +Finished Epoch[1]: [Training Set] Train Loss Per Sample = 0.62975813 EvalErr Per Sample = 0.62975813 Ave Learn Rate Per Sample = 0.1000000015 Epoch Time=5250.689 +Finished Epoch[1]: [Validation Set] Train Loss Per Sample = 0.2035009 EvalErr Per Sample = 0.2035009 + +------ code changed and the following need to be verified ---- +------ May 29 2015 +-------------------------------------------------------------- Network Training Examples: ========================== * rnnlu.config diff --git a/ExampleSetups/SLU/rnnlu.config b/ExampleSetups/SLU/rnnlu.config index c53df1cca..e2278233d 100644 --- a/ExampleSetups/SLU/rnnlu.config +++ b/ExampleSetups/SLU/rnnlu.config @@ -110,7 +110,7 @@ LSTM=[ randomize=None # number of utterances to be allocated for each minibatch - nbruttsineachrecurrentiter=1 + nbruttsineachrecurrentiter=10 # if writerType is set, we will cache to a binary file # if the binary file exists, we will use it instead of parsing this file @@ -210,7 +210,9 @@ LSTM=[ # if writerType is set, we will cache to a binary file # if the binary file exists, we will use it instead of parsing this file # writerType=BinaryReader - + + nbruttsineachrecurrentiter=10 + #### write definition wfile=$ExpDir$\sequenceSentence.valid.bin #wsize - inital size of the file in MB