#GEN -*-*-*-*-*-*-*-*-*-*-*- general info -*-*-*-*-*-*-*-*-*-*-*- Method : MLP::NeuralNetwork TMVA Release : 3.8.13 [198669] ROOT Release : 5.18/00 [332288] Creator : lista Date : Mon Feb 9 12:48:27 2009 Host : Linux lxbuild063.cern.ch 2.6.9-78.0.1.EL.cernsmp #1 SMP Tue Aug 5 11:01:13 CEST 2008 i686 i686 i386 GNU/Linux Dir : /data1/home/lista/stat/03-TMVA Training events: 200000 #OPT -*-*-*-*-*-*-*-*-*-*-*-*- options -*-*-*-*-*-*-*-*-*-*-*-*- # Set by User: Normalise: "True" [Normalise input variables] NCycles: "200" [Number of training cycles] HiddenLayers: "N+2,N" [Specification of hidden layer architecture] NeuronInputType: "sum" [Neuron input function type] LearningRate: "0.02" [ANN learning rate parameter] DecayRate: "0.01" [Decay rate for learning parameter] # Default: D: "False" [use-decorrelated-variables flag (depreciated)] VarTransform: "None" [Variable transformation method] VarTransformType: "Signal" [Use signal or background events for var transform] NbinsMVAPdf: "60" [Number of bins used to create MVA PDF] NsmoothMVAPdf: "2" [Number of smoothing iterations for MVA PDF] V: "False" [Verbose mode] VerboseLevel: "Info" [Verbosity level] H: "False" [Print classifier-specific help message] CreateMVAPdfs: "False" [Create PDFs for classifier outputs] TxtWeightFilesOnly: "True" [if True, write all weights as text files] NeuronType: "tanh" [Neuron activation function type] TrainingMethod: "BP" [Train with Back-Propagation (BP - default) or Genetic Algorithm (GA - slower and worse)] TestRate: "10" [Test for overtraining performed at each #th epochs] BPMode: "sequential" [Back-propagation learning mode: sequential or batch] BatchSize: "-1" [Batch size: number of events/batch, only set if in Batch Mode, -1 for BatchSize=number_of_events] ## #VAR -*-*-*-*-*-*-*-*-*-*-*-* variables *-*-*-*-*-*-*-*-*-*-*-*- NVar 2 x x 'F' [-25.9051780701,28.9163246155] y y 'F' [-10.0186662674,11.3229227066] #MAT -*-*-*-*-*-*-*-*-* transformation data -*-*-*-*-*-*-*-*-*- #WGT -*-*-*-*-*-*-*-*-*-*-*-*- weights -*-*-*-*-*-*-*-*-*-*-*-*- Weights (layer0,neuron0)-(layer1,neuron0): -0.213524276775 (layer0,neuron0)-(layer1,neuron1): 3.66016148829 (layer0,neuron0)-(layer1,neuron2): 6.94936422634 (layer0,neuron0)-(layer1,neuron3): -0.873220328299 (layer0,neuron1)-(layer1,neuron0): -5.64740983296 (layer0,neuron1)-(layer1,neuron1): -2.52130036188 (layer0,neuron1)-(layer1,neuron2): -2.09138202692 (layer0,neuron1)-(layer1,neuron3): 6.23634121296 (layer0,neuron2)-(layer1,neuron0): -2.86931384972 (layer0,neuron2)-(layer1,neuron1): 0.176818000339 (layer0,neuron2)-(layer1,neuron2): -4.4619554297 (layer0,neuron2)-(layer1,neuron3): -2.1920753237 (layer1,neuron0)-(layer2,neuron0): -3.24202661857 (layer1,neuron0)-(layer2,neuron1): -6.42336280668 (layer1,neuron1)-(layer2,neuron0): 3.5589540949 (layer1,neuron1)-(layer2,neuron1): 0.211259999298 (layer1,neuron2)-(layer2,neuron0): -2.39318300064 (layer1,neuron2)-(layer2,neuron1): 2.57236328003 (layer1,neuron3)-(layer2,neuron0): 3.12561144474 (layer1,neuron3)-(layer2,neuron1): 0.116945694087 (layer1,neuron4)-(layer2,neuron0): -2.96275660967 (layer1,neuron4)-(layer2,neuron1): -3.85859065675 (layer2,neuron0)-(layer3,neuron0): 1.02187175199 (layer2,neuron1)-(layer3,neuron0): -0.106384209855 (layer2,neuron2)-(layer3,neuron0): -0.0220788944598