Enable ipywidgets¶
To be able to visualize decision trees and DNN weight map, you must enable ipywidgets. To do so, run the following cell, and refresh the page!
!jupyter nbextension enable --py widgetsnbextension
import ROOT
from ROOT import TFile, TMVA, TCut
Enable JS visualization¶
To use new interactive features in notebook we have to enable a module called JsMVA. This can be done by using ipython magic: %jsmva.
%jsmva on
Declaration of Factory¶
First let's start with the classical version of declaration. If you know how to use TMVA in C++ then you can use that version here in python: first we need to pass a string called job name, as second argument we need to pass an opened output TFile (this is optional, if it's present then it will be used to store output histograms) and as third (or second) argument we pass a string which contains all the settings related to Factory (separated with ':' character).
C++ like declaration¶
outputFile = TFile( "TMVA.root", 'RECREATE' )
TMVA.Tools.Instance();
factory = TMVA.Factory( "TMVAClassification", outputFile #this is optional
,"!V:Color:DrawProgressBar:Transformations=I;D;P;G,D:AnalysisType=Classification" )
The options string can contain the following options:
Option | Default | Predefined values | Description |
---|---|---|---|
V | False | - | Verbose flag |
Color | True | - | Flag for colored output |
Transformations | "" | - | List of transformations to test. For example with "I;D;P;U;G" string identity, decorrelation, PCA, uniform and Gaussian transformations will be applied |
Silent | False | - | Batch mode: boolean silent flag inhibiting any output from TMVA after the creation of the factory class object |
DrawProgressBar | True | - | Draw progress bar to display training, testing and evaluation schedule (default: True) |
AnalysisType | Auto | Classification, Regression, Multiclass, Auto | Set the analysis type |
Pythonic version¶
By enabling JsMVA we have new, more readable ways to do the declaration (this applies to all functions, not just to the constructor).
factory = TMVA.Factory("TMVAClassification", TargetFile=outputFile,
V=False, Color=True, DrawProgressBar=True, Transformations=["I", "D", "P", "G", "D"],
AnalysisType="Classification")
Arguments of constructor: The options string can contain the following options:
Keyword | Can be used as positional argument | Default | Predefined values | Description |
---|---|---|---|---|
JobName | yes, 1. | not optional | - | Name of job |
TargetFile | yes, 2. | if not passed histograms won't be saved | - | File to write control and performance histograms histograms |
V | no | False | - | Verbose flag |
Color | no | True | - | Flag for colored output |
Transformations | no | "" | - | List of transformations to test. For example with "I;D;P;U;G" string identity, decorrelation, PCA, uniform and Gaussian transformations will be applied |
Silent | no | False | - | Batch mode: boolean silent flag inhibiting any output from TMVA after the creation of the factory class object |
DrawProgressBar | no | True | - | Draw progress bar to display training, testing and evaluation schedule (default: True) |
AnalysisType | no | Auto | Classification, Regression, Multiclass, Auto | Set the analysis type |
Declaring the DataLoader, adding variables and setting up the dataset¶
First we need to declare a DataLoader and add the variables (passing the variable names used in the test and train trees in input dataset). To add variable names to DataLoader we use the AddVariable function. Arguments of this function:
String containing the variable name. Using ":=" we can add definition too.
String (label to variable, if not present the variable name will be used) or character (defining the type of data points)
If we have label for variable, the data point type still can be passed as third argument
dataset = "tmva_class_example" #the dataset name
loader = TMVA.DataLoader(dataset)
loader.AddVariable( "myvar1 := var1+var2", 'F' )
loader.AddVariable( "myvar2 := var1-var2", "Expression 2", 'F' )
loader.AddVariable( "var3", "Variable 3", 'F' )
loader.AddVariable( "var4", "Variable 4", 'F' )
It is possible to define spectator variables, which are part of the input data set, but which are not used in the MVA training, test nor during the evaluation, but can be used for correlation tests or others. Parameters:
- String containing the definition of spectator variable.
- Label for spectator variable.
- Data type
loader.AddSpectator( "spec1:=var1*2", "Spectator 1", 'F' )
loader.AddSpectator( "spec2:=var1*3", "Spectator 2", 'F' )
After adding the variables we have to add the datas to DataLoader. In order to do this we check if the dataset file doesn't exist in files directory we download from CERN's server. When we have the root file we open it and get the signal and background trees.
if ROOT.gSystem.AccessPathName( "tmva_class_example.root" ) != 0:
ROOT.gSystem.Exec( "wget https://root.cern.ch/files/tmva_class_example.root")
input = TFile.Open( "tmva_class_example.root" )
# Get the signal and background trees for training
signal = input.Get( "TreeS" )
background = input.Get( "TreeB" )
To pass the signal and background trees to DataLoader we use the AddSignalTree and AddBackgroundTree functions, and we set up the corresponding DataLoader variable's too. Arguments of functions:
- Signal/Background tree
- Global weight used in all events in the tree.
# Global event weights (see below for setting event-wise weights)
signalWeight = 1.0
backgroundWeight = 1.0
loader.AddSignalTree(signal, signalWeight)
loader.AddBackgroundTree(background, backgroundWeight)
loader.fSignalWeight = signalWeight
loader.fBackgroundWeight = backgroundWeight
loader.fTreeS = signal
loader.fTreeB = background
With using DataLoader.PrepareTrainingAndTestTree function we apply cuts on input events. In C++ this function also needs to add the options as a string (as we seen in Factory constructor) which with JsMVA can be passed (same as Factory constructor case) as keyword arguments.
Arguments of PrepareTrainingAndTestTree:
Keyword | Can be used as positional argument | Default | Predefined values | Description |
---|---|---|---|---|
SigCut | yes, 1. | - | - | TCut object for signal cut |
Bkg | yes, 2. | - | - | TCut object for background cut |
SplitMode | no | Random | Random, Alternate, Block | Method of picking training and testing events |
MixMode | no | SameAsSplitMode | SameAsSplitMode, Random, Alternate, Block | Method of mixing events of differnt classes into one dataset |
SplitSeed | no | 100 | - | Seed for random event shuffling |
NormMode | no | EqualNumEvents | None, NumEvents, EqualNumEvents | Overall renormalisation of event-by-event weights used in the training (NumEvents: average weight of 1 per event, independently for signal and background; EqualNumEvents: average weight of 1 per event for signal, and sum of weights for background equal to sum of weights for signal) |
nTrain_Signal | no | 0 (all) | - | Number of training events of class Signal |
nTest_Signal | no | 0 (all) | - | Number of test events of class Signal |
nTrain_Background | no | 0 (all) | - | Number of training events of class Background |
nTest_Background | no | 0 (all) | - | Number of test events of class Background |
V | no | False | - | Verbosity |
VerboseLevel | no | Info | Debug, Verbose, Info | Verbosity level |
mycuts = TCut("")
mycutb = TCut("")
loader.PrepareTrainingAndTestTree(SigCut=mycuts, BkgCut=mycutb,
nTrain_Signal=0, nTrain_Background=0, SplitMode="Random", NormMode="NumEvents", V=False)
Visualizing input variables¶
loader.DrawInputVariable("myvar1")
We can also visualize transformations on input variables¶
loader.DrawInputVariable("myvar1", processTrfs=["D", "N"]) #Transformations: I;N;D;P;U;G,D
Correlation matrix of input variables¶
loader.DrawCorrelationMatrix("Signal")
Booking methods¶
To add which we want to train on dataset we have to use the Factory.BookMethod function. This method will add a method and it's options to Factory.
Arguments:
Keyword | Can be used as positional argument | Default | Predefined values | Description |
---|---|---|---|---|
DataLoader | yes, 1. | - | - | Pointer to DataLoader object |
Method | yes, 2. | - | kVariable kCuts , kLikelihood , kPDERS , kHMatrix , kFisher , kKNN , kCFMlpANN , kTMlpANN , kBDT , kDT , kRuleFit , kSVM , kMLP , kBayesClassifier, kFDA , kBoost , kPDEFoam , kLD , kPlugins , kCategory , kDNN , kPyRandomForest , kPyAdaBoost , kPyGTB , kC50 , kRSNNS , kRSVM , kRXGB , kMaxMethod | Selected method number, method numbers defined in TMVA.Types |
MethodTitle | yes, 3. | - | - | Label for method |
* | no | - | - | Other named arguments which are the options for selected method. |
factory.BookMethod( DataLoader=loader, Method=TMVA.Types.kSVM, MethodTitle="SVM",
Gamma=0.25, Tol=0.001, VarTransform="Norm" )
factory.BookMethod( loader,TMVA.Types.kMLP, "MLP",
H=False, V=False, NeuronType="tanh", VarTransform="N", NCycles=600, HiddenLayers="N+5",
TestRate=5, UseRegulator=False )
factory.BookMethod( loader,TMVA.Types.kLD, "LD",
H=False, V=False, VarTransform="None", CreateMVAPdfs=True, PDFInterpolMVAPdf="Spline2",
NbinsMVAPdf=50, NsmoothMVAPdf=10 )
factory.BookMethod( loader,TMVA.Types.kLikelihood,"Likelihood","NSmoothSig[0]=20:NSmoothBkg[0]=20:NSmoothBkg[1]=10",
NSmooth=1, NAvEvtPerBin=50, H=True, V=False,TransformOutput=True,PDFInterpol="Spline2")
factory.BookMethod( loader, TMVA.Types.kBDT, "BDT",
H=False, V=False, NTrees=850, MinNodeSize="2.5%", MaxDepth=3, BoostType="AdaBoost", AdaBoostBeta=0.5,
UseBaggedBoost=True, BaggedSampleFraction=0.5, SeparationType="GiniIndex", nCuts=20 )
Booking DNN: 2 ways (don't use both in the same time)¶
There is two way to book DNN:
- The visual way: run the next cell, and design the network graphically and then click on "Save Network"
factory.BookDNN(loader)
- Classical way
trainingStrategy = [{
"LearningRate": 1e-1,
"Momentum": 0.0,
"Repetitions": 1,
"ConvergenceSteps": 300,
"BatchSize": 20,
"TestRepetitions": 15,
"WeightDecay": 0.001,
"Regularization": "NONE",
"DropConfig": "0.0+0.5+0.5+0.5",
"DropRepetitions": 1,
"Multithreading": True
}, {
"LearningRate": 1e-2,
"Momentum": 0.5,
"Repetitions": 1,
"ConvergenceSteps": 300,
"BatchSize": 30,
"TestRepetitions": 7,
"WeightDecay": 0.001,
"Regularization": "L2",
"DropConfig": "0.0+0.1+0.1+0.1",
"DropRepetitions": 1,
"Multithreading": True
}, {
"LearningRate": 1e-2,
"Momentum": 0.3,
"Repetitions": 1,
"ConvergenceSteps": 300,
"BatchSize": 40,
"TestRepetitions": 7,
"WeightDecay": 0.001,
"Regularization": "L2",
"Multithreading": True
},{
"LearningRate": 1e-3,
"Momentum": 0.1,
"Repetitions": 1,
"ConvergenceSteps": 200,
"BatchSize": 70,
"TestRepetitions": 7,
"WeightDecay": 0.001,
"Regularization": "NONE",
"Multithreading": True
}, {
"LearningRate": 1e-3,
"Momentum": 0.1,
"Repetitions": 1,
"ConvergenceSteps": 200,
"BatchSize": 70,
"TestRepetitions": 7,
"WeightDecay": 0.001,
"Regularization": "NONE",
"Multithreading": True
}]
factory.BookMethod(DataLoader=loader, Method=TMVA.Types.kDNN, MethodTitle="DNN",
H = False, V=False, VarTransform="Normalize", ErrorStrategy="CROSSENTROPY",
Layout=["TANH|100", "TANH|50", "TANH|10", "LINEAR"],
TrainingStrategy=trainingStrategy,Architecture="STANDARD")
Train Methods¶
When you use the jsmva magic, the original C++ version of Factory::TrainAllMethods is rewritten by a new training method, which will produce notebook compatible output during the training, so we can trace the process (progress bar, error plot). For some methods (MLP, DNN, BDT) there will be created a tracer plot (for MLP, DNN test and training error vs epoch, for BDT error fraction and boost weight vs tree number). There are also some method which doesn't support interactive tracing, so for these methods just a simple text will be printed, just to we know that TrainAllMethods function is training this method currently.
For methods where is possible to trace the training interactively there is a stop button, which can stop the training process. This button just stops the training of the current method, and doesn't stop the TrainAllMethods completely.
factory.TrainAllMethods()
Test end evaluate the methods¶
To test test the methods and evaluate the performance we need to run Factory.TestAllMethods and Factory.EvaluateAllMethods functions.
factory.TestAllMethods()
factory.EvaluateAllMethods()
Classifier Output Distributions¶
To draw the classifier output distribution we have to use Factory.DrawOutputDistribution function which is inserted by invoking jsmva magic. The parameters of the function are the following: The options string can contain the following options:
Keyword | Can be used as positional argument | Default | Predefined values | Description |
---|---|---|---|---|
datasetName | yes, 1. | - | - | The name of dataset |
methodName | yes, 2. | - | - | The name of method |
factory.DrawOutputDistribution(dataset, "MLP")
Classifier Probability Distributions¶
To draw the classifier probability distribution we have to use Factory.DrawProbabilityDistribution function which is inserted by invoking jsmva magic. The parameters of the function are the following: The options string can contain the following options:
Keyword | Can be used as positional argument | Default | Predefined values | Description |
---|---|---|---|---|
datasetName | yes, 1. | - | - | The name of dataset |
factory.DrawProbabilityDistribution(dataset, "LD")