As input data is used a toy-MC sample consisting of four Gaussian-distributed and linearly correlated input variables. The methods to be used can be switched on and off by means of booleans, or via the prompt command, for example:
(note that the backslashes are mandatory) If no method given, a default set of classifiers is used. The output file "TMVA.root" can be analysed with the use of dedicated macros (simply say: root -l <macro.C>), which can be conveniently invoked through a GUI that will appear at the end of the run of this macro. Launch the GUI via the command:
Processing /mnt/build/workspace/root-makedoc-v608/rootspi/rdoc/src/v6-08-00-patches/tutorials/tmva/TMVAClassification.C...
==> Start TMVAClassification
--- TMVAClassification : Using input file: ./tmva_class_example.root
DataSetInfo : [dataset] : Added class "Signal"
: Add Tree TreeS of type Signal with 6000 events
DataSetInfo : [dataset] : Added class "Background"
: Add Tree TreeB of type Background with 6000 events
Factory : Booking method: [1mCuts[0m
:
: Use optimization method: "Monte Carlo"
: Use efficiency computation method: "Event Selection"
: Use "FSmart" cuts for variable: 'myvar1'
: Use "FSmart" cuts for variable: 'myvar2'
: Use "FSmart" cuts for variable: 'var3'
: Use "FSmart" cuts for variable: 'var4'
Factory : Booking method: [1mCutsD[0m
:
CutsD : [dataset] : Create Transformation "Decorrelate" with events from all classes.
:
: Transformation, Variable selection :
: Input : variable 'myvar1' <---> Output : variable 'myvar1'
: Input : variable 'myvar2' <---> Output : variable 'myvar2'
: Input : variable 'var3' <---> Output : variable 'var3'
: Input : variable 'var4' <---> Output : variable 'var4'
: Use optimization method: "Monte Carlo"
: Use efficiency computation method: "Event Selection"
: Use "FSmart" cuts for variable: 'myvar1'
: Use "FSmart" cuts for variable: 'myvar2'
: Use "FSmart" cuts for variable: 'var3'
: Use "FSmart" cuts for variable: 'var4'
Factory : Booking method: [1mLikelihood[0m
:
Factory : Booking method: [1mLikelihoodPCA[0m
:
LikelihoodPCA : [dataset] : Create Transformation "PCA" with events from all classes.
:
: Transformation, Variable selection :
: Input : variable 'myvar1' <---> Output : variable 'myvar1'
: Input : variable 'myvar2' <---> Output : variable 'myvar2'
: Input : variable 'var3' <---> Output : variable 'var3'
: Input : variable 'var4' <---> Output : variable 'var4'
Factory : Booking method: [1mPDERS[0m
:
Factory : Booking method: [1mPDEFoam[0m
:
Factory : Booking method: [1mKNN[0m
:
Factory : Booking method: [1mLD[0m
:
DataSetFactory : [dataset] : Number of events in input trees
:
:
: Number of training and testing events
: ---------------------------------------------------------------------------
: Signal -- training events : 1000
: Signal -- testing events : 5000
: Signal -- training and testing events: 6000
: Background -- training events : 1000
: Background -- testing events : 5000
: Background -- training and testing events: 6000
:
DataSetInfo : Correlation matrix (Signal):
: ----------------------------------------------
: var1+var2 var1-var2 var3 var4
: var1+var2: +1.000 -0.039 +0.778 +0.931
: var1-var2: -0.039 +1.000 -0.111 +0.033
: var3: +0.778 -0.111 +1.000 +0.860
: var4: +0.931 +0.033 +0.860 +1.000
: ----------------------------------------------
DataSetInfo : Correlation matrix (Background):
: ----------------------------------------------
: var1+var2 var1-var2 var3 var4
: var1+var2: +1.000 +0.032 +0.784 +0.931
: var1-var2: +0.032 +1.000 -0.014 +0.111
: var3: +0.784 -0.014 +1.000 +0.863
: var4: +0.931 +0.111 +0.863 +1.000
: ----------------------------------------------
DataSetFactory : [dataset] :
:
Factory : Booking method: [1mFDA_GA[0m
:
: Create parameter interval for parameter 0 : [-1,1]
: Create parameter interval for parameter 1 : [-10,10]
: Create parameter interval for parameter 2 : [-10,10]
: Create parameter interval for parameter 3 : [-10,10]
: Create parameter interval for parameter 4 : [-10,10]
: User-defined formula string : "(0)+(1)*x0+(2)*x1+(3)*x2+(4)*x3"
: TFormula-compatible formula string: "[0]+[1]*[5]+[2]*[6]+[3]*[7]+[4]*[8]"
Factory : Booking method: [1mMLPBNN[0m
:
MLPBNN : [dataset] : Create Transformation "N" with events from all classes.
:
: Transformation, Variable selection :
: Input : variable 'myvar1' <---> Output : variable 'myvar1'
: Input : variable 'myvar2' <---> Output : variable 'myvar2'
: Input : variable 'var3' <---> Output : variable 'var3'
: Input : variable 'var4' <---> Output : variable 'var4'
MLPBNN : Building Network.
: Initializing weights
Factory : Booking method: [1mSVM[0m
:
SVM : [dataset] : Create Transformation "Norm" with events from all classes.
:
: Transformation, Variable selection :
: Input : variable 'myvar1' <---> Output : variable 'myvar1'
: Input : variable 'myvar2' <---> Output : variable 'myvar2'
: Input : variable 'var3' <---> Output : variable 'var3'
: Input : variable 'var4' <---> Output : variable 'var4'
Factory : Booking method: [1mBDT[0m
:
Factory : Booking method: [1mRuleFit[0m
:
Factory : [1mTrain all methods[0m
Factory : [dataset] : Create Transformation "I" with events from all classes.
:
: Transformation, Variable selection :
: Input : variable 'myvar1' <---> Output : variable 'myvar1'
: Input : variable 'myvar2' <---> Output : variable 'myvar2'
: Input : variable 'var3' <---> Output : variable 'var3'
: Input : variable 'var4' <---> Output : variable 'var4'
Factory : [dataset] : Create Transformation "D" with events from all classes.
:
: Transformation, Variable selection :
: Input : variable 'myvar1' <---> Output : variable 'myvar1'
: Input : variable 'myvar2' <---> Output : variable 'myvar2'
: Input : variable 'var3' <---> Output : variable 'var3'
: Input : variable 'var4' <---> Output : variable 'var4'
Factory : [dataset] : Create Transformation "P" with events from all classes.
:
: Transformation, Variable selection :
: Input : variable 'myvar1' <---> Output : variable 'myvar1'
: Input : variable 'myvar2' <---> Output : variable 'myvar2'
: Input : variable 'var3' <---> Output : variable 'var3'
: Input : variable 'var4' <---> Output : variable 'var4'
Factory : [dataset] : Create Transformation "G" with events from all classes.
:
: Transformation, Variable selection :
: Input : variable 'myvar1' <---> Output : variable 'myvar1'
: Input : variable 'myvar2' <---> Output : variable 'myvar2'
: Input : variable 'var3' <---> Output : variable 'var3'
: Input : variable 'var4' <---> Output : variable 'var4'
Factory : [dataset] : Create Transformation "D" with events from all classes.
:
: Transformation, Variable selection :
: Input : variable 'myvar1' <---> Output : variable 'myvar1'
: Input : variable 'myvar2' <---> Output : variable 'myvar2'
: Input : variable 'var3' <---> Output : variable 'var3'
: Input : variable 'var4' <---> Output : variable 'var4'
TFHandler_Factory : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: 0.015467 1.7574 [ -8.1442 7.2697 ]
: myvar2: -0.032440 1.1075 [ -3.9664 4.0259 ]
: var3: 0.0077507 1.0833 [ -5.0373 4.2785 ]
: var4: 0.15186 1.2738 [ -5.9505 4.6404 ]
: -----------------------------------------------------------
: Preparing the Decorrelation transformation...
TFHandler_Factory : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: -0.10647 1.0000 [ -4.3326 4.1397 ]
: myvar2: -0.048178 1.0000 [ -3.6769 3.6584 ]
: var3: -0.11237 1.0000 [ -3.8235 3.8765 ]
: var4: 0.31310 1.0000 [ -4.0754 3.2826 ]
: -----------------------------------------------------------
: Preparing the Principle Component (PCA) transformation...
TFHandler_Factory : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: -0.23314 2.3249 [ -11.198 8.9858 ]
: myvar2: 0.019827 1.1139 [ -4.0542 3.9417 ]
: var3: -0.0063033 0.59141 [ -2.0115 1.9675 ]
: var4: 0.0014833 0.33924 [ -0.99775 1.0218 ]
: -----------------------------------------------------------
: Preparing the Gaussian transformation...
: Preparing the Decorrelation transformation...
TFHandler_Factory : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: -0.036139 1.0000 [ -3.5532 7.9087 ]
: myvar2: 0.018981 1.0000 [ -3.2726 5.5734 ]
: var3: 0.053788 1.0000 [ -3.3619 7.7733 ]
: var4: -0.029204 1.0000 [ -3.4334 5.2857 ]
: -----------------------------------------------------------
: Ranking input variables (method unspecific)...
IdTransformation : Ranking result (top variable is best ranked)
: -------------------------------------
: Rank : Variable : Separation
: -------------------------------------
: 1 : Variable 4 : 2.790e-01
: 2 : Variable 3 : 1.458e-01
: 3 : myvar1 : 8.452e-02
: 4 : Expression 2 : 2.852e-02
: -------------------------------------
Factory : Train method: Cuts for Classification
:
FitterBase : <MCFitter> Sampling, please be patient ...
: Elapsed time: [1;31m3.74 sec[0m
: ------------------------------------------
Cuts : Cut values for requested signal efficiency: 0.1
: Corresponding background efficiency : 0.00632442
: Transformation applied to input variables : None
: ------------------------------------------
: Cut[ 0]: -5.92112 < myvar1 <= 1e+30
: Cut[ 1]: -1e+30 < myvar2 <= 0.372788
: Cut[ 2]: -1.03103 < var3 <= 1e+30
: Cut[ 3]: 1.92872 < var4 <= 1e+30
: ------------------------------------------
: ------------------------------------------
Cuts : Cut values for requested signal efficiency: 0.2
: Corresponding background efficiency : 0.0195583
: Transformation applied to input variables : None
: ------------------------------------------
: Cut[ 0]: -2.39694 < myvar1 <= 1e+30
: Cut[ 1]: -1e+30 < myvar2 <= 0.10647
: Cut[ 2]: -3.38371 < var3 <= 1e+30
: Cut[ 3]: 1.25336 < var4 <= 1e+30
: ------------------------------------------
: ------------------------------------------
Cuts : Cut values for requested signal efficiency: 0.3
: Corresponding background efficiency : 0.036122
: Transformation applied to input variables : None
: ------------------------------------------
: Cut[ 0]: -7.99661 < myvar1 <= 1e+30
: Cut[ 1]: -1e+30 < myvar2 <= 0.516878
: Cut[ 2]: -2.4921 < var3 <= 1e+30
: Cut[ 3]: 1.01508 < var4 <= 1e+30
: ------------------------------------------
: ------------------------------------------
Cuts : Cut values for requested signal efficiency: 0.4
: Corresponding background efficiency : 0.0682383
: Transformation applied to input variables : None
: ------------------------------------------
: Cut[ 0]: -7.92486 < myvar1 <= 1e+30
: Cut[ 1]: -1e+30 < myvar2 <= 2.14952
: Cut[ 2]: -2.18407 < var3 <= 1e+30
: Cut[ 3]: 1.06279 < var4 <= 1e+30
: ------------------------------------------
: ------------------------------------------
Cuts : Cut values for requested signal efficiency: 0.5
: Corresponding background efficiency : 0.114033
: Transformation applied to input variables : None
: ------------------------------------------
: Cut[ 0]: -5.79159 < myvar1 <= 1e+30
: Cut[ 1]: -1e+30 < myvar2 <= 2.13777
: Cut[ 2]: -3.31357 < var3 <= 1e+30
: Cut[ 3]: 0.773468 < var4 <= 1e+30
: ------------------------------------------
: ------------------------------------------
Cuts : Cut values for requested signal efficiency: 0.6
: Corresponding background efficiency : 0.175228
: Transformation applied to input variables : None
: ------------------------------------------
: Cut[ 0]: -4.44635 < myvar1 <= 1e+30
: Cut[ 1]: -1e+30 < myvar2 <= 1.39479
: Cut[ 2]: -1.08281 < var3 <= 1e+30
: Cut[ 3]: 0.343334 < var4 <= 1e+30
: ------------------------------------------
: ------------------------------------------
Cuts : Cut values for requested signal efficiency: 0.7
: Corresponding background efficiency : 0.250904
: Transformation applied to input variables : None
: ------------------------------------------
: Cut[ 0]: -4.12819 < myvar1 <= 1e+30
: Cut[ 1]: -1e+30 < myvar2 <= 3.1382
: Cut[ 2]: -2.7095 < var3 <= 1e+30
: Cut[ 3]: 0.2305 < var4 <= 1e+30
: ------------------------------------------
: ------------------------------------------
Cuts : Cut values for requested signal efficiency: 0.8
: Corresponding background efficiency : 0.381613
: Transformation applied to input variables : None
: ------------------------------------------
: Cut[ 0]: -3.65399 < myvar1 <= 1e+30
: Cut[ 1]: -1e+30 < myvar2 <= 4.05488
: Cut[ 2]: -2.42379 < var3 <= 1e+30
: Cut[ 3]: -0.08696 < var4 <= 1e+30
: ------------------------------------------
: ------------------------------------------
Cuts : Cut values for requested signal efficiency: 0.9
: Corresponding background efficiency : 0.501033
: Transformation applied to input variables : None
: ------------------------------------------
: Cut[ 0]: -7.95481 < myvar1 <= 1e+30
: Cut[ 1]: -1e+30 < myvar2 <= 2.87258
: Cut[ 2]: -3.51779 < var3 <= 1e+30
: Cut[ 3]: -0.564128 < var4 <= 1e+30
: ------------------------------------------
: Elapsed time for training with 2000 events: [1;31m3.74 sec[0m
Cuts : [dataset] : Evaluation of Cuts on training sample (2000 events)
: Elapsed time for evaluation of 2000 events: [1;31m0.0159 sec[0m
: Creating xml weight file: [0;36mdataset/weights/TMVAClassification_Cuts.weights.xml[0m
: Creating standalone class: [0;36mdataset/weights/TMVAClassification_Cuts.class.C[0m
: TMVA.root:/dataset/Method_Cuts/Cuts
Factory : Training finished
:
Factory : Train method: CutsD for Classification
:
: Preparing the Decorrelation transformation...
TFHandler_CutsD : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: -0.10647 1.0000 [ -4.3326 4.1397 ]
: myvar2: -0.048178 1.0000 [ -3.6769 3.6584 ]
: var3: -0.11237 1.0000 [ -3.8235 3.8765 ]
: var4: 0.31310 1.0000 [ -4.0754 3.2826 ]
: -----------------------------------------------------------
TFHandler_CutsD : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: -0.10647 1.0000 [ -4.3326 4.1397 ]
: myvar2: -0.048178 1.0000 [ -3.6769 3.6584 ]
: var3: -0.11237 1.0000 [ -3.8235 3.8765 ]
: var4: 0.31310 1.0000 [ -4.0754 3.2826 ]
: -----------------------------------------------------------
FitterBase : <MCFitter> Sampling, please be patient ...
: Elapsed time: [1;31m2.78 sec[0m
: ------------------------------------------------------------------------------------------------------------------------
CutsD : Cut values for requested signal efficiency: 0.1
: Corresponding background efficiency : 0
: Transformation applied to input variables : "Deco"
: ------------------------------------------------------------------------------------------------------------------------
: Cut[ 0]: -1e+30 < + 1.1153*[myvar1] + 0.035656*[myvar2] - 0.19996*[var3] - 0.79684*[var4] <= -0.776575
: Cut[ 1]: -1e+30 < + 0.035656*[myvar1] + 0.91607*[myvar2] + 0.11642*[var3] - 0.13113*[var4] <= 2.87483
: Cut[ 2]: -0.420335 < - 0.19996*[myvar1] + 0.11642*[myvar2] + 1.7869*[var3] - 0.78589*[var4] <= 1e+30
: Cut[ 3]: 0.851491 < - 0.79684*[myvar1] - 0.13113*[myvar2] - 0.78589*[var3] + 2.155*[var4] <= 1e+30
: ------------------------------------------------------------------------------------------------------------------------
: ------------------------------------------------------------------------------------------------------------------------
CutsD : Cut values for requested signal efficiency: 0.2
: Corresponding background efficiency : 0.00122458
: Transformation applied to input variables : "Deco"
: ------------------------------------------------------------------------------------------------------------------------
: Cut[ 0]: -1e+30 < + 1.1153*[myvar1] + 0.035656*[myvar2] - 0.19996*[var3] - 0.79684*[var4] <= 0.396766
: Cut[ 1]: -1e+30 < + 0.035656*[myvar1] + 0.91607*[myvar2] + 0.11642*[var3] - 0.13113*[var4] <= 0.275176
: Cut[ 2]: -2.02873 < - 0.19996*[myvar1] + 0.11642*[myvar2] + 1.7869*[var3] - 0.78589*[var4] <= 1e+30
: Cut[ 3]: 1.10575 < - 0.79684*[myvar1] - 0.13113*[myvar2] - 0.78589*[var3] + 2.155*[var4] <= 1e+30
: ------------------------------------------------------------------------------------------------------------------------
: ------------------------------------------------------------------------------------------------------------------------
CutsD : Cut values for requested signal efficiency: 0.3
: Corresponding background efficiency : 0.00423588
: Transformation applied to input variables : "Deco"
: ------------------------------------------------------------------------------------------------------------------------
: Cut[ 0]: -1e+30 < + 1.1153*[myvar1] + 0.035656*[myvar2] - 0.19996*[var3] - 0.79684*[var4] <= 3.08016
: Cut[ 1]: -1e+30 < + 0.035656*[myvar1] + 0.91607*[myvar2] + 0.11642*[var3] - 0.13113*[var4] <= 0.714855
: Cut[ 2]: -1.89111 < - 0.19996*[myvar1] + 0.11642*[myvar2] + 1.7869*[var3] - 0.78589*[var4] <= 1e+30
: Cut[ 3]: 1.21039 < - 0.79684*[myvar1] - 0.13113*[myvar2] - 0.78589*[var3] + 2.155*[var4] <= 1e+30
: ------------------------------------------------------------------------------------------------------------------------
: ------------------------------------------------------------------------------------------------------------------------
CutsD : Cut values for requested signal efficiency: 0.4
: Corresponding background efficiency : 0.0104513
: Transformation applied to input variables : "Deco"
: ------------------------------------------------------------------------------------------------------------------------
: Cut[ 0]: -1e+30 < + 1.1153*[myvar1] + 0.035656*[myvar2] - 0.19996*[var3] - 0.79684*[var4] <= 1.90636
: Cut[ 1]: -1e+30 < + 0.035656*[myvar1] + 0.91607*[myvar2] + 0.11642*[var3] - 0.13113*[var4] <= 1.54225
: Cut[ 2]: -2.16041 < - 0.19996*[myvar1] + 0.11642*[myvar2] + 1.7869*[var3] - 0.78589*[var4] <= 1e+30
: Cut[ 3]: 1.14096 < - 0.79684*[myvar1] - 0.13113*[myvar2] - 0.78589*[var3] + 2.155*[var4] <= 1e+30
: ------------------------------------------------------------------------------------------------------------------------
: ------------------------------------------------------------------------------------------------------------------------
CutsD : Cut values for requested signal efficiency: 0.5
: Corresponding background efficiency : 0.0191782
: Transformation applied to input variables : "Deco"
: ------------------------------------------------------------------------------------------------------------------------
: Cut[ 0]: -1e+30 < + 1.1153*[myvar1] + 0.035656*[myvar2] - 0.19996*[var3] - 0.79684*[var4] <= 3.43619
: Cut[ 1]: -1e+30 < + 0.035656*[myvar1] + 0.91607*[myvar2] + 0.11642*[var3] - 0.13113*[var4] <= 1.91064
: Cut[ 2]: -1.7203 < - 0.19996*[myvar1] + 0.11642*[myvar2] + 1.7869*[var3] - 0.78589*[var4] <= 1e+30
: Cut[ 3]: 0.98712 < - 0.79684*[myvar1] - 0.13113*[myvar2] - 0.78589*[var3] + 2.155*[var4] <= 1e+30
: ------------------------------------------------------------------------------------------------------------------------
: ------------------------------------------------------------------------------------------------------------------------
CutsD : Cut values for requested signal efficiency: 0.6
: Corresponding background efficiency : 0.0359533
: Transformation applied to input variables : "Deco"
: ------------------------------------------------------------------------------------------------------------------------
: Cut[ 0]: -1e+30 < + 1.1153*[myvar1] + 0.035656*[myvar2] - 0.19996*[var3] - 0.79684*[var4] <= 2.38751
: Cut[ 1]: -1e+30 < + 0.035656*[myvar1] + 0.91607*[myvar2] + 0.11642*[var3] - 0.13113*[var4] <= 3.08498
: Cut[ 2]: -2.28981 < - 0.19996*[myvar1] + 0.11642*[myvar2] + 1.7869*[var3] - 0.78589*[var4] <= 1e+30
: Cut[ 3]: 0.869049 < - 0.79684*[myvar1] - 0.13113*[myvar2] - 0.78589*[var3] + 2.155*[var4] <= 1e+30
: ------------------------------------------------------------------------------------------------------------------------
: ------------------------------------------------------------------------------------------------------------------------
CutsD : Cut values for requested signal efficiency: 0.7
: Corresponding background efficiency : 0.0668718
: Transformation applied to input variables : "Deco"
: ------------------------------------------------------------------------------------------------------------------------
: Cut[ 0]: -1e+30 < + 1.1153*[myvar1] + 0.035656*[myvar2] - 0.19996*[var3] - 0.79684*[var4] <= 3.6647
: Cut[ 1]: -1e+30 < + 0.035656*[myvar1] + 0.91607*[myvar2] + 0.11642*[var3] - 0.13113*[var4] <= 2.57117
: Cut[ 2]: -2.05772 < - 0.19996*[myvar1] + 0.11642*[myvar2] + 1.7869*[var3] - 0.78589*[var4] <= 1e+30
: Cut[ 3]: 0.60765 < - 0.79684*[myvar1] - 0.13113*[myvar2] - 0.78589*[var3] + 2.155*[var4] <= 1e+30
: ------------------------------------------------------------------------------------------------------------------------
: ------------------------------------------------------------------------------------------------------------------------
CutsD : Cut values for requested signal efficiency: 0.8
: Corresponding background efficiency : 0.110299
: Transformation applied to input variables : "Deco"
: ------------------------------------------------------------------------------------------------------------------------
: Cut[ 0]: -1e+30 < + 1.1153*[myvar1] + 0.035656*[myvar2] - 0.19996*[var3] - 0.79684*[var4] <= 2.55808
: Cut[ 1]: -1e+30 < + 0.035656*[myvar1] + 0.91607*[myvar2] + 0.11642*[var3] - 0.13113*[var4] <= 2.76353
: Cut[ 2]: -2.14693 < - 0.19996*[myvar1] + 0.11642*[myvar2] + 1.7869*[var3] - 0.78589*[var4] <= 1e+30
: Cut[ 3]: 0.420759 < - 0.79684*[myvar1] - 0.13113*[myvar2] - 0.78589*[var3] + 2.155*[var4] <= 1e+30
: ------------------------------------------------------------------------------------------------------------------------
: ------------------------------------------------------------------------------------------------------------------------
CutsD : Cut values for requested signal efficiency: 0.9
: Corresponding background efficiency : 0.255091
: Transformation applied to input variables : "Deco"
: ------------------------------------------------------------------------------------------------------------------------
: Cut[ 0]: -1e+30 < + 1.1153*[myvar1] + 0.035656*[myvar2] - 0.19996*[var3] - 0.79684*[var4] <= 3.23653
: Cut[ 1]: -1e+30 < + 0.035656*[myvar1] + 0.91607*[myvar2] + 0.11642*[var3] - 0.13113*[var4] <= 3.06894
: Cut[ 2]: -3.65126 < - 0.19996*[myvar1] + 0.11642*[myvar2] + 1.7869*[var3] - 0.78589*[var4] <= 1e+30
: Cut[ 3]: 0.0991743 < - 0.79684*[myvar1] - 0.13113*[myvar2] - 0.78589*[var3] + 2.155*[var4] <= 1e+30
: ------------------------------------------------------------------------------------------------------------------------
: Elapsed time for training with 2000 events: [1;31m2.78 sec[0m
CutsD : [dataset] : Evaluation of CutsD on training sample (2000 events)
: Elapsed time for evaluation of 2000 events: [1;31m0.0246 sec[0m
: Creating xml weight file: [0;36mdataset/weights/TMVAClassification_CutsD.weights.xml[0m
: Creating standalone class: [0;36mdataset/weights/TMVAClassification_CutsD.class.C[0m
: TMVA.root:/dataset/Method_CutsD/CutsD
Factory : Training finished
:
Factory : Train method: Likelihood for Classification
:
:
: [1m================================================================[0m
: [1mH e l p f o r M V A m e t h o d [ Likelihood ] :[0m
:
: [1m--- Short description:[0m
:
: The maximum-likelihood classifier models the data with probability
: density functions (PDF) reproducing the signal and background
: distributions of the input variables. Correlations among the
: variables are ignored.
:
: [1m--- Performance optimisation:[0m
:
: Required for good performance are decorrelated input variables
: (PCA transformation via the option "VarTransform=Decorrelate"
: may be tried). Irreducible non-linear correlations may be reduced
: by precombining strongly correlated input variables, or by simply
: removing one of the variables.
:
: [1m--- Performance tuning via configuration options:[0m
:
: High fidelity PDF estimates are mandatory, i.e., sufficient training
: statistics is required to populate the tails of the distributions
: It would be a surprise if the default Spline or KDE kernel parameters
: provide a satisfying fit to the data. The user is advised to properly
: tune the events per bin and smooth options in the spline cases
: individually per variable. If the KDE kernel is used, the adaptive
: Gaussian kernel may lead to artefacts, so please always also try
: the non-adaptive one.
:
: All tuning parameters must be adjusted individually for each input
: variable!
:
: <Suppress this message by specifying "!H" in the booking option>
: [1m================================================================[0m
:
: Filling reference histograms
: Building PDF out of reference histograms
: Elapsed time for training with 2000 events: [1;31m0.0227 sec[0m
Likelihood : [dataset] : Evaluation of Likelihood on training sample (2000 events)
: Elapsed time for evaluation of 2000 events: [1;31m0.0127 sec[0m
: Creating xml weight file: [0;36mdataset/weights/TMVAClassification_Likelihood.weights.xml[0m
: Creating standalone class: [0;36mdataset/weights/TMVAClassification_Likelihood.class.C[0m
: TMVA.root:/dataset/Method_Likelihood/Likelihood
Factory : Training finished
:
Factory : Train method: LikelihoodPCA for Classification
:
: Preparing the Principle Component (PCA) transformation...
TFHandler_LikelihoodPCA : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: -0.23314 2.3249 [ -11.198 8.9858 ]
: myvar2: 0.019827 1.1139 [ -4.0542 3.9417 ]
: var3: -0.0063033 0.59141 [ -2.0115 1.9675 ]
: var4: 0.0014833 0.33924 [ -0.99775 1.0218 ]
: -----------------------------------------------------------
: Filling reference histograms
: Building PDF out of reference histograms
: Elapsed time for training with 2000 events: [1;31m0.0182 sec[0m
LikelihoodPCA : [dataset] : Evaluation of LikelihoodPCA on training sample (2000 events)
: Elapsed time for evaluation of 2000 events: [1;31m0.0218 sec[0m
: Creating xml weight file: [0;36mdataset/weights/TMVAClassification_LikelihoodPCA.weights.xml[0m
: Creating standalone class: [0;36mdataset/weights/TMVAClassification_LikelihoodPCA.class.C[0m
: TMVA.root:/dataset/Method_LikelihoodPCA/LikelihoodPCA
Factory : Training finished
:
Factory : Train method: PDERS for Classification
:
: Elapsed time for training with 2000 events: [1;31m0.0036 sec[0m
PDERS : [dataset] : Evaluation of PDERS on training sample (2000 events)
: Elapsed time for evaluation of 2000 events: [1;31m0.364 sec[0m
: Creating xml weight file: [0;36mdataset/weights/TMVAClassification_PDERS.weights.xml[0m
: Creating standalone class: [0;36mdataset/weights/TMVAClassification_PDERS.class.C[0m
Factory : Training finished
:
Factory : Train method: PDEFoam for Classification
:
PDEFoam : NormMode=NUMEVENTS chosen. Note that only NormMode=EqualNumEvents ensures that Discriminant values correspond to signal probabilities.
: Build up discriminator foam
: Elapsed time: [1;31m0.343 sec[0m
: Elapsed time for training with 2000 events: [1;31m0.385 sec[0m
PDEFoam : [dataset] : Evaluation of PDEFoam on training sample (2000 events)
: Elapsed time for evaluation of 2000 events: [1;31m0.033 sec[0m
: Creating xml weight file: [0;36mdataset/weights/TMVAClassification_PDEFoam.weights.xml[0m
: writing foam DiscrFoam to file
: Foams written to file: [0;36mdataset/weights/TMVAClassification_PDEFoam.weights_foams.root[0m
: Creating standalone class: [0;36mdataset/weights/TMVAClassification_PDEFoam.class.C[0m
Factory : Training finished
:
Factory : Train method: KNN for Classification
:
:
: [1m================================================================[0m
: [1mH e l p f o r M V A m e t h o d [ KNN ] :[0m
:
: [1m--- Short description:[0m
:
: The k-nearest neighbor (k-NN) algorithm is a multi-dimensional classification
: and regression algorithm. Similarly to other TMVA algorithms, k-NN uses a set of
: training events for which a classification category/regression target is known.
: The k-NN method compares a test event to all training events using a distance
: function, which is an Euclidean distance in a space defined by the input variables.
: The k-NN method, as implemented in TMVA, uses a kd-tree algorithm to perform a
: quick search for the k events with shortest distance to the test event. The method
: returns a fraction of signal events among the k neighbors. It is recommended
: that a histogram which stores the k-NN decision variable is binned with k+1 bins
: between 0 and 1.
:
: [1m--- Performance tuning via configuration options: [0m
:
: The k-NN method estimates a density of signal and background events in a
: neighborhood around the test event. The method assumes that the density of the
: signal and background events is uniform and constant within the neighborhood.
: k is an adjustable parameter and it determines an average size of the
: neighborhood. Small k values (less than 10) are sensitive to statistical
: fluctuations and large (greater than 100) values might not sufficiently capture
: local differences between events in the training set. The speed of the k-NN
: method also increases with larger values of k.
:
: The k-NN method assigns equal weight to all input variables. Different scales
: among the input variables is compensated using ScaleFrac parameter: the input
: variables are scaled so that the widths for central ScaleFrac*100% events are
: equal among all the input variables.
:
: [1m--- Additional configuration options: [0m
:
: The method inclues an option to use a Gaussian kernel to smooth out the k-NN
: response. The kernel re-weights events using a distance to the test event.
:
: <Suppress this message by specifying "!H" in the booking option>
: [1m================================================================[0m
:
KNN : <Train> start...
: Reading 2000 events
: Number of signal events 1000
: Number of background events 1000
: Creating kd-tree with 2000 events
: Computing scale factor for 1d distributions: (ifrac, bottom, top) = (80%, 10%, 90%)
ModulekNN : Optimizing tree for 4 variables with 2000 values
: <Fill> Class 1 has 1000 events
: <Fill> Class 2 has 1000 events
: Elapsed time for training with 2000 events: [1;31m0.00309 sec[0m
KNN : [dataset] : Evaluation of KNN on training sample (2000 events)
: Elapsed time for evaluation of 2000 events: [1;31m0.0483 sec[0m
: Creating xml weight file: [0;36mdataset/weights/TMVAClassification_KNN.weights.xml[0m
: Creating standalone class: [0;36mdataset/weights/TMVAClassification_KNN.class.C[0m
Factory : Training finished
:
Factory : Train method: LD for Classification
:
:
: [1m================================================================[0m
: [1mH e l p f o r M V A m e t h o d [ LD ] :[0m
:
: [1m--- Short description:[0m
:
: Linear discriminants select events by distinguishing the mean
: values of the signal and background distributions in a trans-
: formed variable space where linear correlations are removed.
: The LD implementation here is equivalent to the "Fisher" discriminant
: for classification, but also provides linear regression.
:
: (More precisely: the "linear discriminator" determines
: an axis in the (correlated) hyperspace of the input
: variables such that, when projecting the output classes
: (signal and background) upon this axis, they are pushed
: as far as possible away from each other, while events
: of a same class are confined in a close vicinity. The
: linearity property of this classifier is reflected in the
: metric with which "far apart" and "close vicinity" are
: determined: the covariance matrix of the discriminating
: variable space.)
:
: [1m--- Performance optimisation:[0m
:
: Optimal performance for the linear discriminant is obtained for
: linearly correlated Gaussian-distributed variables. Any deviation
: from this ideal reduces the achievable separation power. In
: particular, no discrimination at all is achieved for a variable
: that has the same sample mean for signal and background, even if
: the shapes of the distributions are very different. Thus, the linear
: discriminant often benefits from a suitable transformation of the
: input variables. For example, if a variable x in [-1,1] has a
: a parabolic signal distributions, and a uniform background
: distributions, their mean value is zero in both cases, leading
: to no separation. The simple transformation x -> |x| renders this
: variable powerful for the use in a linear discriminant.
:
: [1m--- Performance tuning via configuration options:[0m
:
: <None>
:
: <Suppress this message by specifying "!H" in the booking option>
: [1m================================================================[0m
:
LD : Results for LD coefficients:
: -----------------------
: Variable: Coefficient:
: -----------------------
: myvar1: -0.326
: myvar2: -0.080
: var3: -0.195
: var4: +0.758
: (offset): -0.056
: -----------------------
: Elapsed time for training with 2000 events: [1;31m0.00106 sec[0m
LD : [dataset] : Evaluation of LD on training sample (2000 events)
: Elapsed time for evaluation of 2000 events: [1;31m0.00692 sec[0m
: <CreateMVAPdfs> Separation from histogram (PDF): 0.578 (0.000)
: Dataset[dataset] : Evaluation of LD on training sample
: Creating xml weight file: [0;36mdataset/weights/TMVAClassification_LD.weights.xml[0m
: Creating standalone class: [0;36mdataset/weights/TMVAClassification_LD.class.C[0m
Factory : Training finished
:
Factory : Train method: FDA_GA for Classification
:
:
: [1m================================================================[0m
: [1mH e l p f o r M V A m e t h o d [ FDA_GA ] :[0m
:
: [1m--- Short description:[0m
:
: The function discriminant analysis (FDA) is a classifier suitable
: to solve linear or simple nonlinear discrimination problems.
:
: The user provides the desired function with adjustable parameters
: via the configuration option string, and FDA fits the parameters to
: it, requiring the signal (background) function value to be as close
: as possible to 1 (0). Its advantage over the more involved and
: automatic nonlinear discriminators is the simplicity and transparency
: of the discrimination expression. A shortcoming is that FDA will
: underperform for involved problems with complicated, phase space
: dependent nonlinear correlations.
:
: Please consult the Users Guide for the format of the formula string
: and the allowed parameter ranges:
: http://tmva.sourceforge.net/docu/TMVAUsersGuide.pdf
:
: [1m--- Performance optimisation:[0m
:
: The FDA performance depends on the complexity and fidelity of the
: user-defined discriminator function. As a general rule, it should
: be able to reproduce the discrimination power of any linear
: discriminant analysis. To reach into the nonlinear domain, it is
: useful to inspect the correlation profiles of the input variables,
: and add quadratic and higher polynomial terms between variables as
: necessary. Comparison with more involved nonlinear classifiers can
: be used as a guide.
:
: [1m--- Performance tuning via configuration options:[0m
:
: Depending on the function used, the choice of "FitMethod" is
: crucial for getting valuable solutions with FDA. As a guideline it
: is recommended to start with "FitMethod=MINUIT". When more complex
: functions are used where MINUIT does not converge to reasonable
: results, the user should switch to non-gradient FitMethods such
: as GeneticAlgorithm (GA) or Monte Carlo (MC). It might prove to be
: useful to combine GA (or MC) with MINUIT by setting the option
: "Converger=MINUIT". GA (MC) will then set the starting parameters
: for MINUIT such that the basic quality of GA (MC) of finding global
: minima is combined with the efficacy of MINUIT of finding local
: minima.
:
: <Suppress this message by specifying "!H" in the booking option>
: [1m================================================================[0m
:
FitterBase : <GeneticFitter> Optimisation, please be patient ... (inaccurate progress timing for GA)
: Elapsed time: [1;31m16.8 sec[0m
FDA_GA : Results for parameter fit using "GA" fitter:
: -----------------------
: Parameter: Fit result:
: -----------------------
: Par(0): 0.38837
: Par(1): -0.325755
: Par(2): -0.0794257
: Par(3): -0.19562
: Par(4): 0.75856
: -----------------------
: Discriminator expression: "(0)+(1)*x0+(2)*x1+(3)*x2+(4)*x3"
: Value of estimator at minimum: 0.237969
: Elapsed time for training with 2000 events: [1;31m16.9 sec[0m
FDA_GA : [dataset] : Evaluation of FDA_GA on training sample (2000 events)
: Elapsed time for evaluation of 2000 events: [1;31m0.0193 sec[0m
: Creating xml weight file: [0;36mdataset/weights/TMVAClassification_FDA_GA.weights.xml[0m
: Creating standalone class: [0;36mdataset/weights/TMVAClassification_FDA_GA.class.C[0m
Factory : Training finished
:
Factory : Train method: MLPBNN for Classification
:
:
: [1m================================================================[0m
: [1mH e l p f o r M V A m e t h o d [ MLPBNN ] :[0m
:
: [1m--- Short description:[0m
:
: The MLP artificial neural network (ANN) is a traditional feed-
: forward multilayer perceptron impementation. The MLP has a user-
: defined hidden layer architecture, while the number of input (output)
: nodes is determined by the input variables (output classes, i.e.,
: signal and one background).
:
: [1m--- Performance optimisation:[0m
:
: Neural networks are stable and performing for a large variety of
: linear and non-linear classification problems. However, in contrast
: to (e.g.) boosted decision trees, the user is advised to reduce the
: number of input variables that have only little discrimination power.
:
: In the tests we have carried out so far, the MLP and ROOT networks
: (TMlpANN, interfaced via TMVA) performed equally well, with however
: a clear speed advantage for the MLP. The Clermont-Ferrand neural
: net (CFMlpANN) exhibited worse classification performance in these
: tests, which is partly due to the slow convergence of its training
: (at least 10k training cycles are required to achieve approximately
: competitive results).
:
: [1mOvertraining: [0monly the TMlpANN performs an explicit separation of the
: full training sample into independent training and validation samples.
: We have found that in most high-energy physics applications the
: avaliable degrees of freedom (training events) are sufficient to
: constrain the weights of the relatively simple architectures required
: to achieve good performance. Hence no overtraining should occur, and
: the use of validation samples would only reduce the available training
: information. However, if the perrormance on the training sample is
: found to be significantly better than the one found with the inde-
: pendent test sample, caution is needed. The results for these samples
: are printed to standard output at the end of each training job.
:
: [1m--- Performance tuning via configuration options:[0m
:
: The hidden layer architecture for all ANNs is defined by the option
: "HiddenLayers=N+1,N,...", where here the first hidden layer has N+1
: neurons and the second N neurons (and so on), and where N is the number
: of input variables. Excessive numbers of hidden layers should be avoided,
: in favour of more neurons in the first hidden layer.
:
: The number of cycles should be above 500. As said, if the number of
: adjustable weights is small compared to the training sample size,
: using a large number of training samples should not lead to overtraining.
:
: <Suppress this message by specifying "!H" in the booking option>
: [1m================================================================[0m
:
TFHandler_MLPBNN : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: 0.058742 0.22803 [ -1.0000 1.0000 ]
: myvar2: -0.015557 0.27714 [ -1.0000 1.0000 ]
: var3: 0.083122 0.23258 [ -1.0000 1.0000 ]
: var4: 0.15238 0.24054 [ -1.0000 1.0000 ]
: -----------------------------------------------------------
: Training Network
:
: Finalizing handling of Regulator terms, trainE=0.667252 testE=0.739861
: Done with handling of Regulator terms
: Elapsed time for training with 2000 events: [1;31m21.5 sec[0m
MLPBNN : [dataset] : Evaluation of MLPBNN on training sample (2000 events)
: Elapsed time for evaluation of 2000 events: [1;31m0.0214 sec[0m
: Creating xml weight file: [0;36mdataset/weights/TMVAClassification_MLPBNN.weights.xml[0m
: Creating standalone class: [0;36mdataset/weights/TMVAClassification_MLPBNN.class.C[0m
: Write special histos to file: TMVA.root:/dataset/Method_MLPBNN/MLPBNN
Factory : Training finished
:
Factory : Train method: SVM for Classification
:
TFHandler_SVM : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: 0.058742 0.22803 [ -1.0000 1.0000 ]
: myvar2: -0.015557 0.27714 [ -1.0000 1.0000 ]
: var3: 0.083122 0.23258 [ -1.0000 1.0000 ]
: var4: 0.15238 0.24054 [ -1.0000 1.0000 ]
: -----------------------------------------------------------
: Building SVM Working Set...with 2000 event instances
: Elapsed time for Working Set build: [1;31m0.11 sec[0m
: Sorry, no computing time forecast available for SVM, please wait ...
: Elapsed time: [1;31m0.494 sec[0m
: Elapsed time for training with 2000 events: [1;31m0.608 sec[0m
SVM : [dataset] : Evaluation of SVM on training sample (2000 events)
: Elapsed time for evaluation of 2000 events: [1;31m0.155 sec[0m
: Creating xml weight file: [0;36mdataset/weights/TMVAClassification_SVM.weights.xml[0m
: Creating standalone class: [0;36mdataset/weights/TMVAClassification_SVM.class.C[0m
Factory : Training finished
:
Factory : Train method: BDT for Classification
:
BDT : #events: (reweighted) sig: 1000 bkg: 1000
: #events: (unweighted) sig: 1000 bkg: 1000
: Training 850 Decision Trees ... patience please
: Elapsed time for training with 2000 events: [1;31m1.02 sec[0m
BDT : [dataset] : Evaluation of BDT on training sample (2000 events)
: Elapsed time for evaluation of 2000 events: [1;31m0.308 sec[0m
: Creating xml weight file: [0;36mdataset/weights/TMVAClassification_BDT.weights.xml[0m
: Creating standalone class: [0;36mdataset/weights/TMVAClassification_BDT.class.C[0m
: TMVA.root:/dataset/Method_BDT/BDT
Factory : Training finished
:
Factory : Train method: RuleFit for Classification
:
:
: [1m================================================================[0m
: [1mH e l p f o r M V A m e t h o d [ RuleFit ] :[0m
:
: [1m--- Short description:[0m
:
: This method uses a collection of so called rules to create a
: discriminating scoring function. Each rule consists of a series
: of cuts in parameter space. The ensemble of rules are created
: from a forest of decision trees, trained using the training data.
: Each node (apart from the root) corresponds to one rule.
: The scoring function is then obtained by linearly combining
: the rules. A fitting procedure is applied to find the optimum
: set of coefficients. The goal is to find a model with few rules
: but with a strong discriminating power.
:
: [1m--- Performance optimisation:[0m
:
: There are two important considerations to make when optimising:
:
: 1. Topology of the decision tree forest
: 2. Fitting of the coefficients
:
: The maximum complexity of the rules is defined by the size of
: the trees. Large trees will yield many complex rules and capture
: higher order correlations. On the other hand, small trees will
: lead to a smaller ensemble with simple rules, only capable of
: modeling simple structures.
: Several parameters exists for controlling the complexity of the
: rule ensemble.
:
: The fitting procedure searches for a minimum using a gradient
: directed path. Apart from step size and number of steps, the
: evolution of the path is defined by a cut-off parameter, tau.
: This parameter is unknown and depends on the training data.
: A large value will tend to give large weights to a few rules.
: Similarily, a small value will lead to a large set of rules
: with similar weights.
:
: A final point is the model used; rules and/or linear terms.
: For a given training sample, the result may improve by adding
: linear terms. If best performance is optained using only linear
: terms, it is very likely that the Fisher discriminant would be
: a better choice. Ideally the fitting procedure should be able to
: make this choice by giving appropriate weights for either terms.
:
: [1m--- Performance tuning via configuration options:[0m
:
: I. TUNING OF RULE ENSEMBLE:
:
: [1mForestType [0m: Recomended is to use the default "AdaBoost".
: [1mnTrees [0m: More trees leads to more rules but also slow
: performance. With too few trees the risk is
: that the rule ensemble becomes too simple.
: [1mfEventsMin [0m
: [1mfEventsMax [0m: With a lower min, more large trees will be generated
: leading to more complex rules.
: With a higher max, more small trees will be
: generated leading to more simple rules.
: By changing this range, the average complexity
: of the rule ensemble can be controlled.
: [1mRuleMinDist [0m: By increasing the minimum distance between
: rules, fewer and more diverse rules will remain.
: Initially it is a good idea to keep this small
: or zero and let the fitting do the selection of
: rules. In order to reduce the ensemble size,
: the value can then be increased.
:
: II. TUNING OF THE FITTING:
:
: [1mGDPathEveFrac [0m: fraction of events in path evaluation
: Increasing this fraction will improve the path
: finding. However, a too high value will give few
: unique events available for error estimation.
: It is recomended to usethe default = 0.5.
: [1mGDTau [0m: cutoff parameter tau
: By default this value is set to -1.0.
: This means that the cut off parameter is
: automatically estimated. In most cases
: this should be fine. However, you may want
: to fix this value if you already know it
: and want to reduce on training time.
: [1mGDTauPrec [0m: precision of estimated tau
: Increase this precision to find a more
: optimum cut-off parameter.
: [1mGDNStep [0m: number of steps in path search
: If the number of steps is too small, then
: the program will give a warning message.
:
: III. WARNING MESSAGES
:
: [1mRisk(i+1)>=Risk(i) in path[0m
: [1mChaotic behaviour of risk evolution.[0m
: The error rate was still decreasing at the end
: By construction the Risk should always decrease.
: However, if the training sample is too small or
: the model is overtrained, such warnings can
: occur.
: The warnings can safely be ignored if only a
: few (<3) occur. If more warnings are generated,
: the fitting fails.
: A remedy may be to increase the value
: [1mGDValidEveFrac[0m to 1.0 (or a larger value).
: In addition, if [1mGDPathEveFrac[0m is too high
: the same warnings may occur since the events
: used for error estimation are also used for
: path estimation.
: Another possibility is to modify the model -
: See above on tuning the rule ensemble.
:
: [1mThe error rate was still decreasing at the end of the path[0m
: Too few steps in path! Increase [1mGDNSteps[0m.
:
: [1mReached minimum early in the search[0m
: Minimum was found early in the fitting. This
: may indicate that the used step size [1mGDStep[0m.
: was too large. Reduce it and rerun.
: If the results still are not OK, modify the
: model either by modifying the rule ensemble
: or add/remove linear terms
:
: <Suppress this message by specifying "!H" in the booking option>
: [1m================================================================[0m
:
RuleFit : -------------------RULE ENSEMBLE SUMMARY------------------------
: Tree training method : AdaBoost
: Number of events per tree : 2000
: Number of trees : 20
: Number of generated rules : 188
: Idem, after cleanup : 65
: Average number of cuts per rule : 2.71
: Spread in number of cuts per rules : 1.11
: ----------------------------------------------------------------
:
: GD path scan - the scan stops when the max num. of steps is reached or a min is found
: Estimating the cutoff parameter tau. The estimated time is a pessimistic maximum.
: Best path found with tau = 0.0400 after [1;31m1.45 sec[0m
: Fitting model...
<WARNING> : Risk(i+1)>=Risk(i) in path
: Risk(i+1)>=Risk(i) in path
: Risk(i+1)>=Risk(i) in path
:
: Minimisation elapsed time : [1;31m0.117 sec[0m
: ----------------------------------------------------------------
: Found minimum at step 300 with error = 0.545798
: Reason for ending loop: clear minima found
: ----------------------------------------------------------------
<WARNING> : Reached minimum early in the search
: Check results and maybe decrease GDStep size
: Removed 13 out of a total of 65 rules with importance < 0.001
:
: ================================================================
: M o d e l
: ================================================================
RuleFit : Offset (a0) = 1.81844
: ------------------------------------
: Linear model (weights unnormalised)
: ------------------------------------
: Variable : Weights : Importance
: ------------------------------------
: myvar1 : -2.303e-01 : 0.901
: myvar2 : -3.967e-02 : 0.097
: var3 : 7.026e-03 : 0.017
: var4 : 3.519e-01 : 1.000
: ------------------------------------
: Number of rules = 52
: Printing the first 10 rules, ordered in importance.
: Rule 1 : Importance = 0.5224
: Cut 1 : var3 < 0.286
: Cut 2 : var4 < -0.615
: Rule 2 : Importance = 0.3544
: Cut 1 : 0.664 < myvar1
: Cut 2 : var4 < 1.85
: Rule 3 : Importance = 0.3517
: Cut 1 : -0.804 < myvar1
: Rule 4 : Importance = 0.3250
: Cut 1 : var3 < 0.286
: Cut 2 : -0.615 < var4
: Rule 5 : Importance = 0.3101
: Cut 1 : myvar1 < 1.4
: Cut 2 : -0.725 < myvar2
: Cut 3 : var4 < -0.672
: Rule 6 : Importance = 0.2823
: Cut 1 : myvar1 < 2.87
: Cut 2 : var4 < 1.21
: Rule 7 : Importance = 0.2738
: Cut 1 : var4 < 1.11
: Rule 8 : Importance = 0.2264
: Cut 1 : myvar1 < 1.36
: Cut 2 : myvar2 < 0.22
: Cut 3 : var4 < -0.412
: Rule 9 : Importance = 0.2121
: Cut 1 : 0.22 < myvar2
: Rule 10 : Importance = 0.1985
: Cut 1 : -0.0703 < myvar1
: Cut 2 : 0.358 < myvar2
: Skipping the next 42 rules
: ================================================================
:
<WARNING> : No input variable directory found - BUG?
: Elapsed time for training with 2000 events: [1;31m1.61 sec[0m
RuleFit : [dataset] : Evaluation of RuleFit on training sample (2000 events)
: Elapsed time for evaluation of 2000 events: [1;31m0.00817 sec[0m
: Creating xml weight file: [0;36mdataset/weights/TMVAClassification_RuleFit.weights.xml[0m
: Creating standalone class: [0;36mdataset/weights/TMVAClassification_RuleFit.class.C[0m
: TMVA.root:/dataset/Method_RuleFit/RuleFit
Factory : Training finished
:
: Ranking input variables (method specific)...
: No variable ranking supplied by classifier: Cuts
: No variable ranking supplied by classifier: CutsD
Likelihood : Ranking result (top variable is best ranked)
: -------------------------------------
: Rank : Variable : Delta Separation
: -------------------------------------
: 1 : var4 : 4.895e-02
: 2 : myvar1 : 1.429e-02
: 3 : var3 : 5.082e-03
: 4 : myvar2 : -9.248e-03
: -------------------------------------
LikelihoodPCA : Ranking result (top variable is best ranked)
: -------------------------------------
: Rank : Variable : Delta Separation
: -------------------------------------
: 1 : var4 : 3.692e-01
: 2 : myvar1 : 7.230e-02
: 3 : var3 : 3.123e-02
: 4 : myvar2 : -1.272e-03
: -------------------------------------
: No variable ranking supplied by classifier: PDERS
PDEFoam : Ranking result (top variable is best ranked)
: ----------------------------------------
: Rank : Variable : Variable Importance
: ----------------------------------------
: 1 : var4 : 3.036e-01
: 2 : myvar1 : 2.500e-01
: 3 : var3 : 2.500e-01
: 4 : myvar2 : 1.964e-01
: ----------------------------------------
: No variable ranking supplied by classifier: KNN
LD : Ranking result (top variable is best ranked)
: ---------------------------------
: Rank : Variable : Discr. power
: ---------------------------------
: 1 : var4 : 7.576e-01
: 2 : myvar1 : 3.263e-01
: 3 : var3 : 1.948e-01
: 4 : myvar2 : 8.026e-02
: ---------------------------------
: No variable ranking supplied by classifier: FDA_GA
MLPBNN : Ranking result (top variable is best ranked)
: -------------------------------
: Rank : Variable : Importance
: -------------------------------
: 1 : myvar1 : 4.870e-01
: 2 : var4 : 2.041e-01
: 3 : myvar2 : 1.112e-01
: 4 : var3 : 1.041e-02
: -------------------------------
: No variable ranking supplied by classifier: SVM
BDT : Ranking result (top variable is best ranked)
: ----------------------------------------
: Rank : Variable : Variable Importance
: ----------------------------------------
: 1 : var4 : 2.748e-01
: 2 : myvar1 : 2.593e-01
: 3 : var3 : 2.336e-01
: 4 : myvar2 : 2.322e-01
: ----------------------------------------
RuleFit : Ranking result (top variable is best ranked)
: -------------------------------
: Rank : Variable : Importance
: -------------------------------
: 1 : var4 : 1.000e+00
: 2 : myvar1 : 8.310e-01
: 3 : myvar2 : 5.152e-01
: 4 : var3 : 4.272e-01
: -------------------------------
Factory : === Destroy and recreate all methods via weight files for testing ===
:
: Read cuts optimised using sample of MC events
: Reading 100 signal efficiency bins for 4 variables
: Read cuts optimised using sample of MC events
: Reading 100 signal efficiency bins for 4 variables
: signal and background scales: 0.001 0.001
: Read foams from file: [0;36mdataset/weights/TMVAClassification_PDEFoam.weights_foams.root[0m
: Creating kd-tree with 2000 events
: Computing scale factor for 1d distributions: (ifrac, bottom, top) = (80%, 10%, 90%)
ModulekNN : Optimizing tree for 4 variables with 2000 values
: <Fill> Class 1 has 1000 events
: <Fill> Class 2 has 1000 events
: User-defined formula string : "(0)+(1)*x0+(2)*x1+(3)*x2+(4)*x3"
: TFormula-compatible formula string: "[0]+[1]*[5]+[2]*[6]+[3]*[7]+[4]*[8]"
MLPBNN : Building Network.
: Initializing weights
Factory : [1mTest all methods[0m
Factory : Test method: Cuts for Classification performance
:
Cuts : [dataset] : Evaluation of Cuts on testing sample (10000 events)
: Elapsed time for evaluation of 10000 events: [1;31m0.00615 sec[0m
Factory : Test method: CutsD for Classification performance
:
CutsD : [dataset] : Evaluation of CutsD on testing sample (10000 events)
: Elapsed time for evaluation of 10000 events: [1;31m0.0108 sec[0m
Factory : Test method: Likelihood for Classification performance
:
Likelihood : [dataset] : Evaluation of Likelihood on testing sample (10000 events)
: Elapsed time for evaluation of 10000 events: [1;31m0.104 sec[0m
Factory : Test method: LikelihoodPCA for Classification performance
:
LikelihoodPCA : [dataset] : Evaluation of LikelihoodPCA on testing sample (10000 events)
: Elapsed time for evaluation of 10000 events: [1;31m0.122 sec[0m
Factory : Test method: PDERS for Classification performance
:
PDERS : [dataset] : Evaluation of PDERS on testing sample (10000 events)
: Elapsed time for evaluation of 10000 events: [1;31m1.02 sec[0m
Factory : Test method: PDEFoam for Classification performance
:
PDEFoam : [dataset] : Evaluation of PDEFoam on testing sample (10000 events)
: Elapsed time for evaluation of 10000 events: [1;31m0.144 sec[0m
Factory : Test method: KNN for Classification performance
:
KNN : [dataset] : Evaluation of KNN on testing sample (10000 events)
: Elapsed time for evaluation of 10000 events: [1;31m0.211 sec[0m
Factory : Test method: LD for Classification performance
:
LD : [dataset] : Evaluation of LD on testing sample (10000 events)
: Elapsed time for evaluation of 10000 events: [1;31m0.00867 sec[0m
: Dataset[dataset] : Evaluation of LD on testing sample
Factory : Test method: FDA_GA for Classification performance
:
FDA_GA : [dataset] : Evaluation of FDA_GA on testing sample (10000 events)
: Elapsed time for evaluation of 10000 events: [1;31m0.00748 sec[0m
Factory : Test method: MLPBNN for Classification performance
:
MLPBNN : [dataset] : Evaluation of MLPBNN on testing sample (10000 events)
: Elapsed time for evaluation of 10000 events: [1;31m0.473 sec[0m
Factory : Test method: SVM for Classification performance
:
SVM : [dataset] : Evaluation of SVM on testing sample (10000 events)
: Elapsed time for evaluation of 10000 events: [1;31m10.8 sec[0m
Factory : Test method: BDT for Classification performance
:
BDT : [dataset] : Evaluation of BDT on testing sample (10000 events)
: Elapsed time for evaluation of 10000 events: [1;31m2.79 sec[0m
Factory : Test method: RuleFit for Classification performance
:
RuleFit : [dataset] : Evaluation of RuleFit on testing sample (10000 events)
: Elapsed time for evaluation of 10000 events: [1;31m0.0197 sec[0m
Factory : [1mEvaluate all methods[0m
Factory : Evaluate classifier: Cuts
:
<WARNING> : You have asked for histogram MVA_EFF_BvsS which does not seem to exist in *Results* .. better don't use it
<WARNING> : You have asked for histogram EFF_BVSS_TR which does not seem to exist in *Results* .. better don't use it
TFHandler_Cuts : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: 0.20304 1.7143 [ -9.8605 7.9024 ]
: myvar2: -0.048739 1.1049 [ -4.0854 4.0291 ]
: var3: 0.15976 1.0530 [ -5.3563 4.6430 ]
: var4: 0.42794 1.2213 [ -6.9675 5.0307 ]
: -----------------------------------------------------------
Factory : Evaluate classifier: CutsD
:
<WARNING> : You have asked for histogram MVA_EFF_BvsS which does not seem to exist in *Results* .. better don't use it
TFHandler_CutsD : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: -0.14823 0.98425 [ -5.4957 4.9064 ]
: myvar2: -0.074925 0.99280 [ -3.6989 3.6412 ]
: var3: -0.097101 1.0022 [ -4.4664 4.4326 ]
: var4: 0.64124 0.94684 [ -3.6874 3.7410 ]
: -----------------------------------------------------------
<WARNING> : You have asked for histogram EFF_BVSS_TR which does not seem to exist in *Results* .. better don't use it
TFHandler_CutsD : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: -0.10647 1.0000 [ -4.3326 4.1397 ]
: myvar2: -0.048178 1.0000 [ -3.6769 3.6584 ]
: var3: -0.11237 1.0000 [ -3.8235 3.8765 ]
: var4: 0.31310 1.0000 [ -4.0754 3.2826 ]
: -----------------------------------------------------------
TFHandler_CutsD : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: -0.14823 0.98425 [ -5.4957 4.9064 ]
: myvar2: -0.074925 0.99280 [ -3.6989 3.6412 ]
: var3: -0.097101 1.0022 [ -4.4664 4.4326 ]
: var4: 0.64124 0.94684 [ -3.6874 3.7410 ]
: -----------------------------------------------------------
Factory : Evaluate classifier: Likelihood
:
Likelihood : [dataset] : Loop over test events and fill histograms with classifier response...
:
TFHandler_Likelihood : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: 0.20304 1.7143 [ -9.8605 7.9024 ]
: myvar2: -0.048739 1.1049 [ -4.0854 4.0291 ]
: var3: 0.15976 1.0530 [ -5.3563 4.6430 ]
: var4: 0.42794 1.2213 [ -6.9675 5.0307 ]
: -----------------------------------------------------------
Factory : Evaluate classifier: LikelihoodPCA
:
TFHandler_LikelihoodPCA : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: 0.74178 2.2502 [ -12.862 10.364 ]
: myvar2: -0.14040 1.1170 [ -4.0632 3.9888 ]
: var3: -0.18967 0.58222 [ -2.2874 1.9945 ]
: var4: -0.32089 0.33116 [ -1.4056 0.88219 ]
: -----------------------------------------------------------
LikelihoodPCA : [dataset] : Loop over test events and fill histograms with classifier response...
:
TFHandler_LikelihoodPCA : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: 0.74178 2.2502 [ -12.862 10.364 ]
: myvar2: -0.14040 1.1170 [ -4.0632 3.9888 ]
: var3: -0.18967 0.58222 [ -2.2874 1.9945 ]
: var4: -0.32089 0.33116 [ -1.4056 0.88219 ]
: -----------------------------------------------------------
Factory : Evaluate classifier: PDERS
:
PDERS : [dataset] : Loop over test events and fill histograms with classifier response...
:
TFHandler_PDERS : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: 0.20304 1.7143 [ -9.8605 7.9024 ]
: myvar2: -0.048739 1.1049 [ -4.0854 4.0291 ]
: var3: 0.15976 1.0530 [ -5.3563 4.6430 ]
: var4: 0.42794 1.2213 [ -6.9675 5.0307 ]
: -----------------------------------------------------------
Factory : Evaluate classifier: PDEFoam
:
PDEFoam : [dataset] : Loop over test events and fill histograms with classifier response...
:
TFHandler_PDEFoam : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: 0.20304 1.7143 [ -9.8605 7.9024 ]
: myvar2: -0.048739 1.1049 [ -4.0854 4.0291 ]
: var3: 0.15976 1.0530 [ -5.3563 4.6430 ]
: var4: 0.42794 1.2213 [ -6.9675 5.0307 ]
: -----------------------------------------------------------
Factory : Evaluate classifier: KNN
:
KNN : [dataset] : Loop over test events and fill histograms with classifier response...
:
TFHandler_KNN : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: 0.20304 1.7143 [ -9.8605 7.9024 ]
: myvar2: -0.048739 1.1049 [ -4.0854 4.0291 ]
: var3: 0.15976 1.0530 [ -5.3563 4.6430 ]
: var4: 0.42794 1.2213 [ -6.9675 5.0307 ]
: -----------------------------------------------------------
Factory : Evaluate classifier: LD
:
LD : [dataset] : Loop over test events and fill histograms with classifier response...
:
: Also filling probability and rarity histograms (on request)...
TFHandler_LD : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: 0.20304 1.7143 [ -9.8605 7.9024 ]
: myvar2: -0.048739 1.1049 [ -4.0854 4.0291 ]
: var3: 0.15976 1.0530 [ -5.3563 4.6430 ]
: var4: 0.42794 1.2213 [ -6.9675 5.0307 ]
: -----------------------------------------------------------
Factory : Evaluate classifier: FDA_GA
:
FDA_GA : [dataset] : Loop over test events and fill histograms with classifier response...
:
TFHandler_FDA_GA : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: 0.20304 1.7143 [ -9.8605 7.9024 ]
: myvar2: -0.048739 1.1049 [ -4.0854 4.0291 ]
: var3: 0.15976 1.0530 [ -5.3563 4.6430 ]
: var4: 0.42794 1.2213 [ -6.9675 5.0307 ]
: -----------------------------------------------------------
Factory : Evaluate classifier: MLPBNN
:
TFHandler_MLPBNN : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: 0.083080 0.22243 [ -1.2227 1.0821 ]
: myvar2: -0.019635 0.27648 [ -1.0298 1.0008 ]
: var3: 0.11576 0.22607 [ -1.0685 1.0783 ]
: var4: 0.20452 0.23064 [ -1.1921 1.0737 ]
: -----------------------------------------------------------
MLPBNN : [dataset] : Loop over test events and fill histograms with classifier response...
:
TFHandler_MLPBNN : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: 0.083080 0.22243 [ -1.2227 1.0821 ]
: myvar2: -0.019635 0.27648 [ -1.0298 1.0008 ]
: var3: 0.11576 0.22607 [ -1.0685 1.0783 ]
: var4: 0.20452 0.23064 [ -1.1921 1.0737 ]
: -----------------------------------------------------------
Factory : Evaluate classifier: SVM
:
TFHandler_SVM : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: 0.083080 0.22243 [ -1.2227 1.0821 ]
: myvar2: -0.019635 0.27648 [ -1.0298 1.0008 ]
: var3: 0.11576 0.22607 [ -1.0685 1.0783 ]
: var4: 0.20452 0.23064 [ -1.1921 1.0737 ]
: -----------------------------------------------------------
SVM : [dataset] : Loop over test events and fill histograms with classifier response...
:
TFHandler_SVM : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: 0.083080 0.22243 [ -1.2227 1.0821 ]
: myvar2: -0.019635 0.27648 [ -1.0298 1.0008 ]
: var3: 0.11576 0.22607 [ -1.0685 1.0783 ]
: var4: 0.20452 0.23064 [ -1.1921 1.0737 ]
: -----------------------------------------------------------
Factory : Evaluate classifier: BDT
:
BDT : [dataset] : Loop over test events and fill histograms with classifier response...
:
TFHandler_BDT : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: 0.20304 1.7143 [ -9.8605 7.9024 ]
: myvar2: -0.048739 1.1049 [ -4.0854 4.0291 ]
: var3: 0.15976 1.0530 [ -5.3563 4.6430 ]
: var4: 0.42794 1.2213 [ -6.9675 5.0307 ]
: -----------------------------------------------------------
Factory : Evaluate classifier: RuleFit
:
RuleFit : [dataset] : Loop over test events and fill histograms with classifier response...
:
TFHandler_RuleFit : Variable Mean RMS [ Min Max ]
: -----------------------------------------------------------
: myvar1: 0.20304 1.7143 [ -9.8605 7.9024 ]
: myvar2: -0.048739 1.1049 [ -4.0854 4.0291 ]
: var3: 0.15976 1.0530 [ -5.3563 4.6430 ]
: var4: 0.42794 1.2213 [ -6.9675 5.0307 ]
: -----------------------------------------------------------
:
: Evaluation results ranked by best signal efficiency and purity (area)
: -------------------------------------------------------------------------------------------------------------------
: DataSet MVA
: Name: Method: ROC-integ
: dataset LD : 0.837
: dataset FDA_GA : 0.836
: dataset MLPBNN : 0.861
: dataset LikelihoodPCA : 0.791
: dataset SVM : 0.809
: dataset CutsD : 0.901
: dataset BDT : 0.923
: dataset RuleFit : 0.731
: dataset KNN : 0.745
: dataset PDEFoam : 0.710
: dataset PDERS : 0.671
: dataset Cuts : 0.792
: dataset Likelihood : 0.615
: -------------------------------------------------------------------------------------------------------------------
:
: Testing efficiency compared to training efficiency (overtraining check)
: -------------------------------------------------------------------------------------------------------------------
: DataSet MVA Signal efficiency: from test sample (from training sample)
: Name: Method: @B=0.01 @B=0.10 @B=0.30
: -------------------------------------------------------------------------------------------------------------------
: dataset LD : 0.351 (0.426) 0.763 (0.762) 0.922 (0.925)
: dataset FDA_GA : 0.351 (0.416) 0.763 (0.762) 0.922 (0.925)
: dataset MLPBNN : 0.337 (0.412) 0.750 (0.769) 0.917 (0.926)
: dataset LikelihoodPCA : 0.317 (0.370) 0.751 (0.763) 0.913 (0.929)
: dataset SVM : 0.357 (0.369) 0.744 (0.752) 0.912 (0.919)
: dataset CutsD : 0.242 (0.394) 0.727 (0.788) 0.903 (0.922)
: dataset BDT : 0.225 (0.418) 0.640 (0.724) 0.854 (0.904)
: dataset RuleFit : 0.215 (0.258) 0.637 (0.692) 0.865 (0.899)
: dataset KNN : 0.185 (0.257) 0.554 (0.630) 0.819 (0.847)
: dataset PDEFoam : 0.000 (0.192) 0.458 (0.494) 0.775 (0.773)
: dataset PDERS : 0.170 (0.155) 0.466 (0.447) 0.754 (0.733)
: dataset Cuts : 0.108 (0.145) 0.446 (0.479) 0.741 (0.746)
: dataset Likelihood : 0.082 (0.085) 0.385 (0.383) 0.684 (0.661)
: -------------------------------------------------------------------------------------------------------------------
:
Dataset:dataset : Created tree 'TestTree' with 10000 events
:
Dataset:dataset : Created tree 'TrainTree' with 2000 events
:
Factory : [1mThank you for using TMVA![0m
: [1mFor citation information, please visit: http://tmva.sf.net/citeTMVA.html[0m
==> Wrote root file: TMVA.root
==> TMVAClassification is done!
(int) 0