Line: 1 to 1 | ||||||||
---|---|---|---|---|---|---|---|---|
MCLimits Fitting Code: Tool for Measuring Sensitivity | ||||||||
Line: 7 to 7 | ||||||||
This tool is currently utilised at both CDF![]() ![]() ![]() ![]() ![]() | ||||||||
Changed: | ||||||||
< < | MCLimits Code | |||||||
> > | MCLimits Tarball | |||||||
Changed: | ||||||||
< < | The MC Limits tarball should be downloaded following the link above. Several files are included in the tarball in a directory, mclimit/. Details of the files are given below, alongside their purpose.
Prerequisites | |||||||
> > | The MC Limits tarball should be downloaded following the link above. Several files are included in the tarball in a directory, mclimit/. These are detailed below:
Documentation | |||||||
Changed: | ||||||||
< < | Project Aims | |||||||
> > | There is significant documentation which comes along with the tarball of the code. This contains details of the methods and functions available in mclimits as well as giving an outline of the statistical framework
The Code
Test Files - Beware | |||||||
Changed: | ||||||||
< < | This project aims to develop an Artificial Neural Network (ANN) system and fitting software for the analysis of data from inclusive Higgs searches at ATLAS involving a lepton trigger and Higgs decay to b+bbar. It also aims to document this software, both its development and how it is to be used for data analysis - specifically separating signal from background and obtaining exclusion limits based on the anticipated luminosity. It is desirable that the documentation should allow the data analysis system to be used by those without advanced knowledge of ROOT or C++ programming, and should include a user-friendly guide to gaining access to relevant resources (TWikis, Grid, etc.). | |||||||
> > | These files DO NOT work (see the README file for details):
| |||||||
Changed: | ||||||||
< < | Prerequisites | |||||||
> > | This file might work, but hasn't been tested. It takes the output of the mclimits method, csm_model::print() (which is detailed in the documentation, mclimits_csm.pdf) and produces a webpage of template names and systematics sources for each template.
| |||||||
Changed: | ||||||||
< < | Before starting data analysis it is necessary to gain access to several online resources, specifically the Glasgow ATLAS TWiki and the central ATLAS TWiki, as well as obtaining a Grid certificate to allow use of Subversion, a software repository tool. The first and third of these can be accomplished easily enough - see instructions on Full TWiki Access and Grid Computing respectively. However, gaining access to the central ATLAS Twiki![]() | |||||||
> > | Running the Code | |||||||
Changed: | ||||||||
< < | Once you have a Grid certificate you must follow the instructions on the PPEIT TWiki![]() | |||||||
> > | To run the mclimits package, you need to produce a file which prepares the inputs to mclimits. This is what tchan* is an example of, however these examples require some specific root files that don't come along with the tarball. As such, and a simple example is provided here. | |||||||
Changed: | ||||||||
< < | Online Resources | |||||||
> > | Useful Output | |||||||
Changed: | ||||||||
< < | The central ATLAS TWiki is an invaluable resource, with extensive information on all aspects of data analysis, including a detailed guide to the relevant software. For an introductory overview see the analysis and computing workbooks here![]() | |||||||
> > | The outcomes from the MCLimits code is heavily dependent on the methods which are selected and run. However there are a set of standard outputs that are of use when doing a hypothesis test, and assessing sensitivity. These are detailed here. | |||||||
Changed: | ||||||||
< < | Tools | |||||||
> > | - PDFs | |||||||
Changed: | ||||||||
< < | ANN :- This is a kind of algorithm with a structure consisting of "neurons" organised in a sequence of layers. The most common type, which is used here, is the Multi-Layer Perceptron (MLP), which comprises three kinds of layer. The input neurons are activated by a set trigger, and once activated they pass data on to a further set of "hidden" neurons (which can in principle be organised into any number of layers, but most frequently one or two), and finally the processed data is forwarded to the output neurons. | |||||||
> > | - Lumi95 | |||||||
Changed: | ||||||||
< < | The key feature of a neural network is its ability to be "trained" to recognise patterns in data, allowing high efficiency algorithms to be developed with relative ease. This training is typically done with sample data which has been generated artificially, resulting in an algorithm that is very effective at recognising certain patterns in data sets. The only shortcoming is the danger of "over-training" an ANN, meaning that it becomes overly discriminating and searches across a narrower range of patterns than is desired (one countermeasure is to add extra noise to training data). | |||||||
> > | - Lumi3sigma | |||||||
Changed: | ||||||||
< < | ROOT :- This is a data analysis package developed by CERN specifically for use in nuclear and particle physics. It is extremely flexible and powerful, especially for generating graphical representations (e.g. histograms) of data. Essentially, it may be thought of as an object-oriented language compatible with C++, although ROOT code generally does not resemble pure C++. Some useful ROOT resources are: | |||||||
> > | - Lumi5sigma | |||||||
Changed: | ||||||||
< < | ||||||||
> > | - CLs +/- 1,2sigma | |||||||
Changed: | ||||||||
< < |
| |||||||
> > | - 1-CLb +/- 1,2sigma | |||||||
Changed: | ||||||||
< < | ||||||||
> > | - J. Heinrich's Bayesian Limit results (s95med, s95p1,s95p2,s95m1,s95m2) | |||||||
Changed: | ||||||||
< < | Limitations | |||||||
> > | Useful Statistics Documentation | |||||||
Changed: | ||||||||
< < |
TMVA Training PlotsThere is a macro in the latest NNFitter version![]()
Setting up SubversionNote - the first part of this section outlines the procedure used to set up a repository with Subversion. Once the repository has been created, these commands do not need to be used again, so if you are working with code from an existing repository you should start at checking out.Running analysis & making ntuplesThis is the procedure used to remake the ATLAS ntuples, using code from CERN Subversion repositories. These ntuples were then used as input for the neural net. -- CatherineWright - 2010-02-12Online ResourcesThe central ATLAS TWiki is an invaluable resource, with extensive information on all aspects of data analysis, including a detailed guide to the relevant software. For an introductory overview see the analysis and computing workbooks here![]() ToolsANN :- This is a kind of algorithm with a structure consisting of "neurons" organised in a sequence of layers. The most common type, which is used here, is the Multi-Layer Perceptron (MLP), which comprises three kinds of layer. The input neurons are activated by a set trigger, and once activated they pass data on to a further set of "hidden" neurons (which can in principle be organised into any number of layers, but most frequently one or two), and finally the processed data is forwarded to the output neurons. The key feature of a neural network is its ability to be "trained" to recognise patterns in data, allowing high efficiency algorithms to be developed with relative ease. This training is typically done with sample data which has been generated artificially, resulting in an algorithm that is very effective at recognising certain patterns in data sets. The only shortcoming is the danger of "over-training" an ANN, meaning that it becomes overly discriminating and searches across a narrower range of patterns than is desired (one countermeasure is to add extra noise to training data). ROOT :- This is a data analysis package developed by CERN specifically for use in nuclear and particle physics. It is extremely flexible and powerful, especially for generating graphical representations (e.g. histograms) of data. Essentially, it may be thought of as an object-oriented language compatible with C++, although ROOT code generally does not resemble pure C++. Some useful ROOT resources are:
Limitations
TMVA Training PlotsThere is a macro in the latest NNFitter version![]()
Setting up SubversionNote - the first part of this section outlines the procedure used to set up a repository with Subversion. Once the repository has been created, these commands do not need to be used again, so if you are working with code from an existing repository you should start at checking out.Running analysis & making ntuplesThis is the procedure used to remake the ATLAS ntuples, using code from CERN Subversion repositories. These ntuples were then used as input for the neural net. | |||||||
> > | There is a myriad of statistics books and papers available. A simple list of papers that are useful for understanding the statistical concepts adopted in the mclimits code are listed here:
| |||||||
-- CatherineWright - 2010-02-12 |