"""
The Configuration defaults
These are the default values for MAUS and its components. Please DO NOT CHANGE
this unless you want to change this quantity for everybody. Values can be
overridden by setting configuration_file parameter on the comamnd line, for
example
bin/simulate_mice.py --configuration_file my_configuration.py
"""
# This file is part of MAUS: http://micewww.pp.rl.ac.uk:8080/projects/maus
#
# MAUS is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# MAUS is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with MAUS. If not, see .
# NOTE: please use lower_case_and_underscores for new configuration cards
import os
# cuts_parameters holds the parameters that are needed to determine the cuts
recon_cuts = {
"min_tof":28.,
"max_tof":32.,
"min_mom_US":130.,
"max_mom_US":150.,
"min_mom_DS":50.,
"max_mom_DS":100.,
"min_mom_loss":5.,
"max_mom_loss":43.4,
"min_mass":99.0,
"max_mass":110.0,
"good_pval":0.02,
}
primary_correction = { 'scale' : 1.0, 'shift' : -10.0 }
type_of_dataflow = 'pipeline_single_thread'
input_json_file_name = "maus_input.json"
input_json_file_type = "text"
output_json_file_name = "maus_output.json"
output_json_file_type = "text"
input_root_file_name = "maus_input.root"
output_root_file_name = "maus_output.root"
# For InputCppRoot Enter an array of spill numbers here to select them from the
# input root file. Leaving the array empty automatically loads all spills.
# (Note present version emits selected spills for all input run numbers)
# e.g. selected_spills = [ 2, 34, 432, 3464 ]
selected_spills = []
data_maximum_reference_count = 110
# one_big_file - puts everything in output_root_file_name
# one_file_per_run - splits and inserts xxx_.xxx for each run, like
# maus_output_1111.root
# end_of_run_file_per_run - as above, and places in
# //
# users responsibility to ensure that
# exists but MAUS will make subdirectories
output_root_file_mode = "one_big_file"
end_of_run_output_root_directory = os.environ.get("MAUS_WEB_MEDIA_RAW")+"/end_of_run/" \
if (os.environ.get("MAUS_WEB_MEDIA_RAW") != None) else os.getcwd()
# Used, for now, to determine what level of
# c++ log messages are reported to the user:
# 0 = debug info (and std::cout)
# 1 = run info
# 2 = warnings
# 3 = errors (and std::cerr)
# 4 = fatal
# >4 = silent
# Doesnt effect python
verbose_level = 1
# Used for switching the general maus running log on or off
# 0 = off
# 1 = on
log_level = 0
errors_to_stderr = None # None = from verbose_level; else True or False
errors_to_json = True
on_error = 'none' # none, halt or raise
will_do_stack_trace = verbose_level < 1 # set to True to make stack trace on C++
# exception
# set how headers and footers are handled - "append" will set to
# append headers and footers to the output; dont_append will set to not append
# headers and footers to the output. Affects JobHeader, JobFooter, RunHeader and
# RunFooter
#
# For example, if the user wants to copy data from one format to another, he
# should set to dont_append to avoid header and footer information being taken
# to the output
header_and_footer_mode = "append" #append or dont_append
# Dictionary of variable to be set when using G4BL to generate particles. If
# "get_magnet_currents_pa_cdb" is set to True magnet currents & proton absorber
# thickness will be retrieved from the CDB for the run_number specified
g4bl = {"run_number":2873,"q_1":1.066,"q_2":-1.332,"q_3":0.927,"d_1":-1.302,"d_2":-0.396,\
"d_s":3.837,"particles_per_spill":0,"rotation_angle":0,"translation_z":1000.0,\
"protonabsorberin":1,"proton_absorber_thickness":93,"proton_number":1E9,"proton_weight":1,\
"particle_charge":'all',"file_path":'MAUS_ROOT_DIR/src/map/MapPyBeamlineSimulation/G4bl',\
"get_magnet_currents_pa_cdb":False,"random_seed":1,
"seed_algorithm":"random_seed_and_spill_number",
}
# Used by MapPyRemoveTracks.
keep_only_muon_tracks = False
# Used by MapCppSimulation
keep_tracks = False # set to true to keep start and end point of every track
keep_steps = False # set to true to keep start and end point of every track and
# every step point
#simulation_geometry_filename = "Stage4.dat" # geometry used by simulation - should be replaced by CDB version
simulation_geometry_filename = "Test.dat" # geometry used by simulation - should be replaced by CDB version
simulation_geometry_debug = False
check_volume_overlaps = False
maximum_number_of_steps = 50000000 # particles are killed after this number of
# steps (assumed to be stuck in the fields)
simulation_reference_particle = { # used for setting particle phase
"position":{"x":0.0, "y":-0.0, "z":-6400.0},
"momentum":{"x":0.0, "y":0.0, "z":1.0},
"particle_id":-13, "energy":226.0, "time":0.0, "random_seed":10,
"spin":{"x":0.0, "y":0.0, "z":1.0}
# "beam_polarisation" : "Flat"
}
everything_special_virtual = False
# geant4 physics model
physics_model = "QGSP_BERT" # Physics package loaded by MAUS to set default values; modifications can be made
reference_physics_processes = "mean_energy_loss" # controls the physics processes of the reference particle. Set to "none" to disable all physics processes; or "mean_energy_loss" to make the reference particle see deterministic energy loss only
physics_processes = "standard" # controls the physics processes of normal particles. Set to "none" to disable all physics processes; "mean_energy_loss" to enable deterministic energy loss only; or "standard" to enable all physics processes
particle_decay = True # set to true to activate particle decay, or False to inactivate particle decay
polarised_decay = False # set to true to make muons decay with the correct distribution if the beam is polarised; if true, will force spin tracking on (if spin tracking is off, spin vectors will not be propagated to the decay point)
charged_pion_half_life = -1. # set the pi+, pi- half life [ns]. Negative value means use geant4 default
muon_half_life = -1. # set the mu+, mu- half life [ns]. Negative value means use geant4 default
production_threshold = 0.5 # set the threshold for delta ray production [mm]
fine_grained_production_threshold = { # set the production threshold per pid and per Geant4 region; regions are defined in MiceModules
# "region_name":{"11":0.5, "-11":100.}
# "another_region_name":{"11":0.5, "-11":100.}
}
default_keep_or_kill = True
# map of string pdg pids; always keep particles on creation if their pdg maps to True; always kill particles on creation if their pdg maps to False. Default comes from default_keep_or_kill
keep_or_kill_particles = {"mu+":True, "mu-":True,
"nu_e":False, "anti_nu_e":False,
"nu_mu":False, "anti_nu_mu":False,
"nu_tau":False, "anti_nu_tau":False,
}
max_step_length = 100. # default maximum step size during tracking (override with G4StepMax in MiceModule)
max_track_time = 1.e9 # kill tracks with time above max_time (override with G4TimeMax in MiceModule)
max_track_length = 1.e8 # kill tracks with track length above max_track_length (override with G4TrackMax in MiceModule)
kinetic_energy_threshold = 0.1 # kill tracks with initial kinetic energy below energy_threshold (override with G4KinMin in MiceModule)
field_tracker_absolute_error = 1.e-4 # set absolute error on MAUS internal stepping routines - used by e.g. VirtualPlanes to control accuracy of interpolation
field_tracker_relative_error = 1.e-4 # set relative error on MAUS internal stepping routines - used by e.g. VirtualPlanes to control accuracy of interpolation
stepping_algorithm = "ClassicalRK4" # numerical integration routine
spin_tracking = False # set to true to use G4 routines to precess the spin vector as particles go through EM fields
delta_one_step = -1. # Geant4 step accuracy parameter
delta_intersection = -1.
epsilon_min = -1.
epsilon_max = -1.
miss_distance = -1.
maximum_module_depth = 10 # maximum depth for MiceModules as used by the simulation
# geant4 visualisation (not event display)
geant4_visualisation = False
visualisation_viewer = "VRML2FILE" # only supported option
visualisation_theta = 90.
visualisation_phi = 90.
visualisation_zoom = 1.
accumulate_tracks = 0 # this accumulates the tracks into one vrml file 1 = yes, 0 = no
#particle colour alterations
default_vis_colour = {"red":0. , "green":100. ,"blue":0.} #green
pi_plus_vis_colour = {"red":255. , "green":250. ,"blue":240.} # white
pi_minus_vis_colour = {"red":105. , "green":105. ,"blue":105.} # grey
mu_plus_vis_colour = {"red":25. , "green":25. ,"blue":112.} # dark blue
mu_minus_vis_colour = {"red":135. , "green":206. ,"blue":250.} # light blue
e_plus_vis_colour = {"red":250. , "green":0. ,"blue":0.} # red
e_minus_vis_colour = {"red":250. , "green":69. ,"blue":0.} # orange red
gamma_vis_colour = {"red":255. , "green":20. ,"blue":147.} # pink
neutron_vis_colour = {"red":139. , "green":69. ,"blue":19.} # brown
photon_vis_colour = {"red":255. , "green":255. ,"blue":0.} # yellow
# used by InputPySpillGenerator to determine the number of empty spills that
# will be generated by the simulation
spill_generator_number_of_spills = 100
# used by MapPyBeamMaker to generate input particle data
# This is a sample beam distribution based on guesses by Chris Rogers of what
# an optimised beam might look like
beam = {
"particle_generator":"binomial", # routine for generating empty primaries
"binomial_n":50, # number of coin tosses
"binomial_p":0.5, # probability of making a particle on each toss
"random_seed":5, # random seed for beam generation; controls also how the MC
# seeds are generated
# "particle_generator":"g4bl", # Uses G4BL as input for MAUS
# "g4bl_generator":"True", # Call G4BL each time new spill is created
# "random_seed":5,
"definitions":[
##### MUONS #######
{
"reference":simulation_reference_particle, # reference particle
"random_seed_algorithm":"incrementing_random", # algorithm for seeding MC
"weight":90., # probability of generating a particle
"transverse":{
"transverse_mode":"constant_solenoid", # transverse distribution matched to constant solenoid field
"emittance_4d":6., # 4d emittance
"normalised_angular_momentum":0.1, # angular momentum from diffuser
"bz":4.e-3 # magnetic field strength for angular momentum calculation
},
"longitudinal":{"longitudinal_mode":"sawtooth_time", # longitudinal distribution sawtooth in time
"momentum_variable":"p", # Gaussian in total momentum (options energy, pz)
"sigma_p":25., # RMS total momentum
"t_start":-1.e6, # start time of sawtooth
"t_end":+1.e6}, # end time of sawtooth
"coupling":{"coupling_mode":"none"} # no dispersion
#"spin":{"x":0.0, "y":0.0, "z":1.0}
},
##### PIONS #####
{ # as above...
"reference":{
"position":{"x":0.0, "y":-0.0, "z":-6400.0},
"momentum":{"x":0.0, "y":0.0, "z":1.0},
"spin":{"x":0.0, "y":0.0, "z":1.0},
"particle_id":211, "energy":285.0, "time":0.0, "random_seed":10
},
"random_seed_algorithm":"incrementing_random",
"weight":2.,
"transverse":{"transverse_mode":"constant_solenoid", "emittance_4d":6.,
"normalised_angular_momentum":0.1, "bz":4.e-3},
"longitudinal":{"longitudinal_mode":"sawtooth_time",
"momentum_variable":"p",
"sigma_p":25.,
"t_start":-1.e6,
"t_end":+1.e6},
"coupling":{"coupling_mode":"none"}
# "spin":{"x":0.0, "y":0.0, "z":1.0}
},
##### ELECTRONS #####
{ # as above...
"reference":{
"position":{"x":0.0, "y":-0.0, "z":-6400.0},
"momentum":{"x":0.0, "y":0.0, "z":1.0},
"spin":{"x":0.0, "y":0.0, "z":1.0},
"particle_id":-11, "energy":200.0, "time":0.0, "random_seed":10
},
"random_seed_algorithm":"incrementing_random",
"weight":8.,
"transverse":{"transverse_mode":"constant_solenoid", "emittance_4d":6.,
"normalised_angular_momentum":0.1, "bz":4.e-3},
"longitudinal":{"longitudinal_mode":"sawtooth_time",
"momentum_variable":"p",
"sigma_p":25.,
"t_start":-1.e6,
"t_end":+1.e6},
"coupling":{"coupling_mode":"none"}
# "spin":{"x":0.0, "y":0.0, "z":1.0}
}]
}
# this is used by reconstruction; if set to an empty string, automatically
# acquires simulation_geometry_filename
reconstruction_geometry_filename = ""
# scifi tracker digitization
SciFi_BG_VLPCOn = 1
SciFi_BG_VLPCRate = 0.05
SciFiKunoSum = 318.5 # Sum of channel #s in 3 planes if they form a spoint
SciFiKunoSumT1S5 = 320.0 # Sum of channel #s in 3 planes if they form a spoint for T1 S5
SciFiKunoTolerance = 3.0 # Kuno conjecture tolerance
SciFiDigitizationNPECut = 2.0
SciFiDigitizationADCNoiseCut = False
SciFiConfigDir = os.environ.get("MAUS_ROOT_DIR") if (os.environ.get("MAUS_ROOT_DIR") != None) else None
SciFiMappingFileName = "scifi_mapping.txt"
SciFiCalibrationFileName = "scifi_calibration.txt"
SciFiBadChannelsFileName = "scifi_bad_channels.txt"
SciFiCalibMethod = "Current" # Date/Current/Run
SciFiCalibSrc = 7057 # exmple: "Date" - 1984-09-14 00:10:00.0 "Run" - 7057
SciFiMUXNum = 7
SciFiFiberDecayConst = 2.7
SciFiFiberConvFactor = 3047.1
SciFiFiberTrappingEff = 0.056
SciFiFiberMirrorEff = 0.6
SciFiFiberTransmissionEff = 0.8
SciFiMUXTransmissionEff = 1.0
SciFivlpcQE = 0.8
SciFivlpcEnergyRes = 4.0 # MeV
SciFivlpcTimeRes = 0.2 # ns
SciFiadcBits = 8.0
SciFitdcBits = 16
SciFitdcFactor = 1.0
SciFinPlanes = 3
SciFinStations = 5
SciFinTrackers = 2
SciFiCalibrateMC = True
SciFiNPECut = 2.0 # photoelectrons
SciFiClustExcept = 100 # exception is thrown
SciFi_sigma_tracker0_station5 = 0.4298 # Position error associated with station 5 (mm)
SciFi_sigma_triplet = 0.3844 # Position error for stations 1 to 4 (mm)
SciFi_sigma_z = 0.081 # mm
SciFi_sigma_duplet = 0.6197 # mm
SciFi_sigma_phi_1to4 = 1.0
SciFi_sigma_phi_5 = 1.0
SciFiMCSStationError = 2.0 # mm
SciFiClusterReconOn = True
SciFiSpacepointReconOn = True
SciFiPRHelicalTkUSOn = 0 # TkUS helical pattern recognition: 0 = auto, 1 = off, 2 = on
SciFiPRHelicalTkDSOn = 0 # TkDS helical pattern recognition: 0 = auto, 1 = off, 2 = on
SciFiPRStraightTkUSOn = 0 # TkUS straight pattern recognition: 0 = auto, 1 = off, 2 = on
SciFiPRStraightTkDSOn = 0 # TDUS straight pattern recognition: 0 = auto, 1 = off, 2 = on
SciFiPatRecMissingSpSearchOn = False # Do we seach for seed spoints missed by helical fit?
SciFiPatRecMissingSpCut = 2 # Distance (mm) below which a missing spoint should added to a track
SciFiPatRecSErrorMethod = 0 # How to calc error on s, 0 = station res, 1 = error prop
SciFiPatRecVerbosity = 0 # The verbosity of the pat rec (0 - quiet, 1 - more)
SciFiPatRecLongitudinalFitter = 1 # 0 - ntruns and linear s-z fit, 1 - ROOT and MINUIT
SciFiPatRecLineFitter = 0 # Choose the patrec straight line fitter, 0 = custom lsq, 1 = ROOT
SciFiPatRecCircleFitter = 1 # Choose the patrec circle fitter, 0 = custom lsq, 1 = MINUIT
SciFiStraightRoadCut = 7.0 # The road cut in pat rec for straights (mm)
SciFiStraightChi2Cut = 50.0 # Chi^2 on pat rec straight track fit
SciFiHelicalRadiusLimit = 5.0 # The point at which we call it a straight track if we have TOF information.
SciFiRadiusResCut = 150.0 # Helix radius cut (mm) for pattern recognition
SciFiPatRecCircleChi2Cut = 5.0 # Chi^2 on pat rec least squares circle fit
SciFiPatRecCircleMinuitChi2Cut = 10 # Chi^2 on pat rec minuit circle fit
SciFiNTurnsCut = 1.0 # Cut used when resolving number of turns between tracker stations (mm)
SciFiPatRecSZChi2Cut = 150.0 # Chi^2 cut on pat rec s-z fit
SciFiPatRecLongMinuitChi2Cut = 65.0 # Chi^2 cut on pat rec minuit longitudinal fit
SciFiPatRecCircleErrorWeight = 1.0 # Weight to artificially scale the error going to xy fit
SciFiPatRecSZErrorWeight = 1.0 # Weight to artificially scale the error going to sz fit
SciFiMaxPt = 180.0 # Transverse momentum upper limit cut used in pattern recognition
SciFiMinPz = 50.0 # Longitudinal momentum lower limit cut used in pattern recognition
SciFiPatRecDebugOn = False # Set Pattern Recogntition to debug mode
SciFiPatRecDebugFileName = "pattern_recognition_debug.root" # Output file name for patrec debug
SciFiParams_Pitch = 1.4945
SciFiParams_Station_Radius = 160. # Used as cut by SpacePointReconstruction
SciFiParams_RMS = 370.
SciFiMinuitSearchNStepsNormal = 100 # steps used when searching for seed chisq min location
SciFiMinuitSearchNStepsLowRadius = 200 # steps used when searching for seed chisq min location
SciFiMinuitStepSize = 0.01 # minuit step size
SciFiMinuitMaxFunctionCalls = 1000000 # Max calls minuit is allowed when doing a minimisation
SciFiMinuitMaxIterations = 100000 # Max num of iterations minuit is allowed
SciFiMinuitTolerance = 0.001 # minuit tolerance
SciFiMinuitMaxSZChisq = 50.0 # Max chisq a candidate minuit ds/dz value can have
SciFiMinuitDsDzLowerLimit = -0.35 # Minimum allowed value for ds/dz
SciFiMinuitDsDzUpperLimit = 0.35 # Maximum allowed value for ds/dz
SciFiPatRecLowRadius = 10.0 # values below 10mm imply a "low radius" when patrec fits struggle
# Parameters used for MCS and E loss correction
#SciFiParams_Z = 5.61291
#SciFiParams_A = 104.15 # g/mol
SciFiParams_Z = 3.5
SciFiParams_A = 6.5
SciFiParams_Plane_Width = 0.6273 # mm
SciFiParams_Radiation_Length = 413.124 # mm
SciFiParams_Density = 1.06 #g/cm3, 0.00106 g/mm3
SciFiParams_Mean_Excitation_Energy = 68.7 / 1.0E6 # MeV
SciFiParams_Density_Correction = 0.164541
#MylarParams_Z = 5.
#MylarParams_A = 192.2 # g/mol
MylarParams_Z = 8.727
MylarParams_A = 4.545 # g/mol
MylarParams_Plane_Width = 0.025 # mm
MylarParams_Radiation_Length = 285.364 # mm
MylarParams_Density = 1.4 #g/cm3, 0.0014 g/mm3
MylarParams_Mean_Excitation_Energy = 78.7 / 1.0E6 # eV
MylarParams_Density_Correction = 0.126782
GasParams_Z = 2.
GasParams_A = 4. # g/mol
GasParams_Radiation_Length = 5671130. # mm
GasParams_Density = 0.000166322 # 1.66322e-04 g/cm3
GasParams_Mean_Excitation_Energy = 41.8 / 1.0E6 # eV
GasParams_Density_Correction = 0.13443
SciFiTestVirtualTracksStraight = {"rms_position" : 70.0, "rms_angle" : 0.19} # Description of straight tracks to simulate
SciFiTestVirtualTracksHelix = {"rms_position" : 70.0, "rms_pt" : 30.0, "pz" : 200.0} # Description of helical tracks to simulate
SciFiTestVirtualMethod = "virtual" # How to test the scifi recon. Choose from "straight", "helical" or "virtual"
SciFiTestVirtualMakeDigits = False
SciFiTestVirtualMakeClusters = True
SciFiTestVirtualMakeSpacepoints = False
SciFiTestVirtualSmear = 0.431425 # Simulate measurement error on alpha with Gaussian Smearing this is the Std Dev.
# Set the smear value to negative to force a quantisation of alpha - like a real measurement
SciFiPRCorrection = 1.1776
SciFiPRBias = 0.2269
SciFiPRCorrectionsOutputFile = "SciFiMomentumCorrections.root" #File to output momentum correction data to (produced by reducer)
SciFiPRCorrectSeed = 0 # 0 : Don't correct PR seed, 1 : apply corretions from corrections file
SciFiPRCorrectionsFile = "SciFiMomentumCorrections.root" # File to use to apply seed corrections (used in PR Mapper)
SciFiSeedCovariance = 1000.0 # Error estimate for Seed values of the Kalman Fit
SciFiSeedPatRec = True
SciFiKalmanOn = True # Flag to turn on the tracker Kalman Fit
SciFiKalmanCorrectPz = True # Flag to turn on the Kalman Pz correlations
SciFiPatRecOn = True # Flag to turn on the tracker Pattern Recognition
SciFiKalman_use_MCS = True # flag to add MCS to the Kalman Fit
SciFiKalman_use_Eloss = True # flag to add Eloss to the Kalman Fit
SciFiKalmanVerbose = False # Dump information per fitted track
SciFiDefaultMomentum = 200.0 # Default momentum to assume for straight tracks
# How we determine the track ratings:
SciFiExcellentNumTrackpoints = 15
SciFiGoodNumTrackpoints = 10
SciFiPoorNumTrackpoints = 10
SciFiExcellentNumSpacepoints = 5
SciFiGoodNumSpacepoints = 5
SciFiPoorNumSpacepoints = 4
SciFiExcellentPValue = 0.05
SciFiGoodPValue = 0.01
SciFiPoorPValue = 0.0
SciFi_CalibratedNoiseFile = "LEDNoisePlots.root"
SciFi_CalibratedNoiseRateScaling = 1.0
SciFi_CalibratedNoiseNPEScaling = 1.5
# configuration database
cdb_upload_url = "http://cdb.mice.rl.ac.uk/cdb/" # target URL for configuration database uploads TestServer::http://rgma19.pp.rl.ac.uk:8080/cdb/
cdb_download_url = "http://cdb.mice.rl.ac.uk/cdb/" # target URL for configuration database downloads
#cdb_download_url = "http://preprodcdb.mice.rl.ac.uk:8080/cdb/" # target URL for configuration database downloads
cdb_cc_download_url = "" # "http://preprodcdb.mice.rl.ac.uk" # target URL for cooling channel configuration database downloads.
# geometry download
geometry_download_wsdl = "geometry?wsdl" # name of the web service used for downloads
geometry_download_directory = "%s/files/geometry/download" % os.environ.get("MAUS_ROOT_DIR") # name of the local directory where downloads will be placed
geometry_download_by = 'run_number' # choose 'run_number' to download by run number, 'current' to use
# the currently valid geometry or 'id' to use the cdb internal id
# (e.g. if it is desired to access an old version of a particular
# geometry)
geometry_download_beamline_for_run = 0
geometry_download_beamline_tag = ''
geometry_download_coolingchannel_tag = ''
geometry_download_run_number = 7469
geometry_download_id = 160
geometry_download_apply_corrections = True
geometry_download_cleanup = True # set to True to clean up after download
g4_step_max = 5.0 # this is the value which shall be placed in the Mice Modules which have been translated from CAD
geometry_use_active_rotations = False # Changes SciFiGeometryHelper to use active rotations for Mice Moduel only running.
# geometry upload
geometry_upload_wsdl = "geometrySuperMouse?wsdl" # name of the web service used for uploads
geometry_upload_directory = "%s/files/geometry/upload" % os.environ.get("MAUS_ROOT_DIR") # name of the local directory where uploads are drawn from
geometry_upload_note = "" # note, pushed to the server to describe the geometry. A note must be specified here (default will throw an exception).
geometry_upload_valid_from = "" # date-time in format like: that the specified installation was made in the experiment. A date-time must be specified here (default will throw an exception).
technical_drawing_name = "" #name and version of the technical drawing from which the CAD model came from.
geometry_upload_cleanup = True # set to True to clean up after upload
technical_drawing_name = "" #name and version of the technical drawing from which the CAD model came from.
#dates need to get geomtry ids
get_ids_start_time = "2012-05-08 09:00:00"
get_ids_stop_time = "2012-05-22 15:47:34.856000"
get_ids_create_file = True
#get beamline information
# This executable will give the run numbers of the runs which the CDB has information on.
# The information is the magnet currents, reasons for run and other information which
# is specific to that run. When downloading a geometry by run number the beamline
# information is merged with the geometrical infomation. Options for querying
# beamline information are; 'all_entries' returns a list of all run numbers with beamline info.
# 'run_number' prints whether info is held for this run number or not.
# 'dates' returns a list of run numbers with info during specified time period.
get_beamline_by = "all_entries"
get_beamline_run_number = ""
get_beamline_start_time = ""
get_beamline_stop_time = ""
# File Numbers
# This following section gives the files numbers of each detector. The numbers speficy the technical drawing
# number for the sphere which represents each detector in the CAD model. These tags are seen in the style sheet
# and are used to replace the location sphere with the detecor geometry whether it is legacy or GDML.
tof_0_file_number = "Iges_10"
tof_1_file_number = "Iges_11"
tof_2_file_number = "Iges_13"
ckov1_file_number = "Iges_19"
ckov2_file_number = "Iges_21"
kl_file_number = "Iges_14"
emr_file_number = "Iges_15"
tracker0_file_number = "Iges_17"
tracker1_file_number = "Iges_18"
absorber0_file_number = "9999"
absorber1_file_number = "Iges_16"
# absorber1_file_number = "9999"
absorber2_file_number = "9999"
# Survey fit information
survey_measurement_record = ""
# This file should include position references and true locations of each detector.
survey_reference_position = ""
use_gdml_source = True
# Survey targets
survey_target_detectors = []
# geometry substitutions for cooling channel current scales
# SSU and SSD rescaled according to C.Rogers
geometry_coolingchannel_scales = {
"SSU-E2": 1.02,
"SSU-C": 1.02,
"SSU-E1": 1.02,
"SSU-M2": 1.0,
"SSU-M1": 1.0,
"FCU-C": 1.0,
"SSD-E2": 1.018,
"SSD-C": 1.018,
"SSD-E1": 1.018,
"SSD-M1": 1.0,
"SSD-M2": 1.0
}
# this is used by ImputCppRealData
Number_of_DAQ_Events = -1
Input_Use_JSON = False
Phys_Events_Only = False
Calib_Events_Only = False
Enable_V1495_Unpacking = False
Enable_EI_Unpacking = True
Enable_V1290_Unpacking = True
Enable_V1731_Unpacking = True
Enable_V1724_Unpacking = True
Enable_V830_Unpacking = True
Enable_VLSB_Unpacking = True
Enable_VLSB_C_Unpacking = False
Enable_DBB_Unpacking = True
Enable_DBBChain_Unpacking = True
Do_V1731_Zero_Suppression = False
V1731_Zero_Suppression_Threshold = 100
Do_V1724_Zero_Suppression = True
V1724_Zero_Suppression_Threshold = 100
Do_VLSB_Zero_Suppression = True
VLSB_Zero_Suppression_Threshold = 1 # in PE count rather than ADC count.
Do_VLSB_C_Zero_Suppression = False
VLSB_C_Zero_Suppression_Threshold = 30
Enable_TOF = True
Enable_EMR = True
Enable_KL = True
Enable_CKOV = True
# DAQ cabling maps
# set the source for cabling to either 'CDB' or 'file'
# if set to 'file' the DAQ_cabling_file card will be used
DAQ_cabling_source = "CDB"
# set the method for retrieving cabling from CDB
# options are: 'run_number' or 'date'
# if DAQ_cabling_by is set to 'date', the DAQ_cabling_date card is used
# Default is to get the map based on run number
DAQ_cabling_by = "run_number"
# date can be 'current' or a date in YYYY-MM-DD hh:mm:ss format
# e.g. DAQ_cabling_date = '2015-06-30 01:02:03 00:00:00'
# or DAQ_cabling_date = 'current'
DAQ_cabling_date = 'current'
# the DAQ_cabling_file card is used only if DAQ_cabling_source above is set to 'file'
DAQ_cabling_file = "/files/cabling/DAQChannelMap.txt"
DAQ_cabling_file_StepI = "/files/cabling/DAQChannelMap_preRun6541.txt"
DAQ_hostname = 'miceraid5'
DAQ_monitor_name = 'MICE_Online_Monitor'
daq_online_file = '' # set to a file name to force InputCppDAQOnlineData to take
# data from a file - mock-up of online for testing, not for
# production use (use offline recon here)
daq_online_spill_delay_time = 0. # delay in seconds between daq reads, intended
# for mocking MICE target pulses
# tof digitization
TOFconversionFactor = 0.005 # MeV
TOFpmtTimeResolution = 0.1 # nanosecond
TOFattenuationLength = 140 * 10 # mm
TOFadcConversionFactor = 0.125
TOFtdcConversionFactor = 0.025 # nanosecond
TOFpmtQuantumEfficiency = 0.25
TOFscintLightSpeed = 170.0 # mm/ns
# KL digitization
KLconversionFactor = 0.000125 # MeV
KLlightCollectionEff = 0.031
KLlightGuideEff = 0.85
KLquantumEff = 0.26
KLlightSpeed = 170.0 # mm/ns
KLattLengthLong = 2400.0 # mm
KLattLengthShort = 200.0 # mm
KLattLengthLongNorm = 0.655 # mm
KLattLengthShortNorm = 0.345 # mm
KLadcConversionFactor = 250000 # nphe/adc
KLpmtGain = 2000000
KLpmtSigmaGain = 1000000
# EMR characteristics
EMRnumberOfPlanes = 48
EMRnumberOfBars = 60 # number of bars in one plane (+ test channel 0)
EMRnBars = 2832 # total number of bars in the EMR
EMRbarLength = 1100 # mm, length of a scintillating bar
EMRbarWidth = 33 # mm, base of the triangle
EMRbarHeight = 17 # mm, height of the triangle
EMRgap = 0.5 # mm, gap between two adjacent bars
# EMR event pre-selection
EMRreferenceDBB = 7 # Reference DBB id for absolute time
EMRprimaryTotUp = 100 # DBB counts, tune down unrealistic signals
#EMRprimaryDeltatLow = -240 # DBB counts, Step I
EMRprimaryDeltatLow = -230 # DBB counts, Step IV
#EMRprimaryDeltatUp = -220 # DBB counts, Step I
EMRprimaryDeltatUp = -210 # DBB counts, Step IV
EMRsecondaryNLow = 2
EMRsecondaryTotLow = 4 # DBB counts
EMRsecondaryBunchingWidth = 10 # ADC counts, 25ns
EMRfom = "median" # figure_Of-Merit for signal calibration
# EMR digitization
EMRdoSampling = 1 # sample as a Poisson distribution
EMRnphPerMeV = 2000 # number of photons per MeV
EMRtrapEff = 0.04 # photon trapping efficiency
EMRseed = 2 # seed state of the pseudorandom number generator, change equivalent to a power cycle of fADC
EMRattenWLSf = 2.0 # dB/m, attentuation factor of the WLS fibres
EMRattenCLRf = 0.35 # dB/m, attentuation factor of the clear fibres
EMRbirksConstant = 0.126 # mm/MeV
EMRsignalEnergyThreshold = 0.8 # MeV
EMRdbbCount = 2.5 # ns, duration of a DBB cycle (f=400MHz)
EMRqeMAPMT = 0.25 # MAPMT quantum efficiency
EMRnadcPerPeMAPMT = 3 # number of ADC counts per photoelectron in the MAPMT
EMRelectronicsResponseSpreadMAPMT = 1 # ADC counts
EMRtimeResponseSpread = 1 # ADC counts
EMRtotFuncP1 = 12.55 # a in a*log(b*Q+c) (Shaping factor)
EMRtotFuncP2 = 0.0252 # b in a*log(b*Q+c) (Scaling factor)
EMRtotFuncP3 = 1.015 # c in a*log(b*Q+c) (Offset factor)
EMRdeltatShift = 12 # ADC counts, distance from the trigger
EMRfadcCount = 2.0 # ns, duration of an fADC cycle (f=500MHz)
#EMRqeSAPMT = 0.11 # SAPMT quantum efficiency, Step I
EMRqeSAPMT = 0.24 # SAPMT quantum efficiency, Step IV
#EMRnadcPerPeSAPMT = 2 # number of ADC counts per photoelectron in the SAPMT, Step I
EMRnadcPerPeSAPMT = 3 # number of ADC counts per photoelectron in the SAPMT, Step IV
EMRelectronicsResponseSpreadSAPMT = 1 # ADC count, Step I
EMRbaselinePosition = 123 # SAPMT signal baseline
EMRbaselineSpread = 10 # SAPMT signal baseline spread
EMRmaximumNoiseLevel = 50 # SAPMT noise maximum value
EMRacquisitionWindow = 302 # ADC counts
EMRsignalIntegrationWindow = 40 # ADC counts
EMRarrivalTimeShift = 40
EMRarrivalTimeSpread = 33
EMRarrivalTimeGausWidth = 3
EMRarrivalTimeUniformWidth = 12.5
EMRpulseShapeLandauWidth = 2
# EMR reconstruction
EMRchargeThreshold = 10 # ADC counts, rejects noise for plane density
EMRpolynomialOrder = 1 # Order of the polynomial to fit the tracks with
EMRmaxMotherDaughterTime = 15000 # ns, ~ 7 decay constants, 99.9% of muons
EMRmaxMotherDaughterDistance = 100 # mm, max distance between mother and daughter
EMRholeFraction = 0.02117 # percentage of the EMR volume that is not PS
# EMR reducer
EMRdensityCut = 0.9
EMRchi2Cut = 2
# this is used by the reconstuction of the TOF detectors
TOF_trigger_station = "tof1"
# this sets the source for the calibrations and cabling
# by default it is from CDB
# set it to 'file' if you want to load local files
# if you set file, then uncomment the calib files below
TOF_calib_source = "CDB"
TOF_cabling_source = "CDB"
#TOF_cabling_file = "/files/cabling/TOFChannelMap.txt"
#TOF_TW_calibration_file = "/files/calibration/tofcalibTW_dec2011.txt"
#TOF_T0_calibration_file = "/files/calibration/tofcalibT0_trTOF1_dec2011.txt"
#TOF_T0_calibration_file = "/files/calibration/tofcalibT0_trTOF0.txt"
#TOF_Trigger_calibration_file = "/files/calibration/tofcalibTrigger_trTOF1_dec2011.txt"
#TOF_Trigger_calibration_file = "/files/calibration/tofcalibTrigger_trTOF0.txt"
# Revert back as it was just for JN muon beams runs in 3.3.4
TOF_findTriggerPixelCut = 0.5 # nanosecond
TOF_makeSpacePointCut = 0.5 # nanosecond
# get cabling and calibrations by either a) run_number or b) date
# default is by run_number
# if set to "date" then set the appropriate TOF_calib_date_from flag below
TOF_calib_by = "run_number"
TOF_cabling_by = "run_number"
# the date for which we want the cabling and calibration
# date can be 'current' or a date in YYYY-MM-DD hh:mm:ss format
TOF_calib_date_from = 'current'
TOF_cabling_date_from = 'current'
Enable_timeWalk_correction = True
Enable_triggerDelay_correction = True
Enable_t0_correction = True
# tof bad channel correction
# give the run range in which the channel is bad
# the correction is hard-coded in TOFSlabHits
TOF_badchan_start = 7200
TOF_badchan_end = 10603
# this is used by the reconstuction of the KL detectors
# this sets the source for the calibrations
# by default it is from CDB
# set it to 'file' if you want to load local files
# if you set file, then uncomment the calib files below
KL_calib_source = "CDB"
KL_calib_date_from = 'current'
# uncomment the KL_calibration_file card below if you set KL_calib_source=file
#KL_calibration_file = "/files/calibration/klcalib.txt"
KL_cabling_file = "/files/cabling/KLChannelMap.txt"
Enable_klgain_correction = True
# this is used by the reconstuction of the EMR detectors
EMR_cabling_file = "/files/cabling/EMRChannelMap.txt"
# this sets the source of the EMR clear fiber length map
EMR_clear_fiber_length_map = "/files/cabling/EMRClearFiberLengthMap.txt"
# this sets the source of the EMR connector attenuation map
EMR_connector_attenuation_map = "/files/cabling/EMRConnectorAttenuationMap.txt"
# this sets the source of the calibrations for the EMR detector
EMR_calib_source = "CDB"
EMR_calib_date_from = 'current'
# EMR_calibration_file card below if you set EMR_calib_source=file
EMR_calib_file = "/files/calibration/emrcalib_cosmics_july2015.txt"
daq_data_path = '%s/src/input/InputCppDAQData' % os.environ.get("MAUS_ROOT_DIR") # path to daq data. Multiple locations can be specified with a space
# daq_data_file = '05466.001' # file name for daq data; if this is just a integer string, MAUS assumes this is a run number. Multiple entries can be specified separated by a space
daq_data_file = '06008.000'
maus_version = "" # set at runtime - do not edit this (changes are ignored)
configuration_file = "" # should be set on the command line only (else ignored)
doc_store_class = "docstore.MongoDBDocumentStore.MongoDBDocumentStore"
doc_collection_name = "spills" # Default document collection name. Only needed if using multi_process mode. If "auto" then a collection name will be auto-generated for spills output by input-transform workflows.
doc_store_event_cache_size = 10**8 # Maximum size of the Mongo cache to cache at any one time in multiprocessing mode, as used by e.g. online code. Corresponds to ~ n/3 spills.
mongodb_host = "localhost" # Default MongoDB host name. Only needed if using MongoDBDocumentStore.
mongodb_port = 27017 # Default MongoDB port. Only needed if using MongoDBDocumentStore.
mongodb_database_name = "mausdb" # Default MongoDB database name. Only needed if using MongoDBDocumentStore.
mongodb_collection_name = "spills" # Default MongoDB collection name. Only needed if using MongoDBDocucmentStore.
# in multiprocessing mode, the timeout after which reconstruction of an event will be abandonded [s]
reconstruction_timeout = 10
# refresh rate for refreshing the plots
reduce_plot_refresh_rate = 10
# Default OutputPyImage image directory. MAUS web application directory.
image_directory = os.environ.get("MAUS_WEB_MEDIA_RAW") if (os.environ.get("MAUS_WEB_MEDIA_RAW") != None) else os.getcwd()
# Default OutputPyImage image directory for end of run data. Will end up as image_directory+"/end_of_run/"
end_of_run_image_directory = ''
# Default OutputPyFile output directory. MAUS web application directory.
output_file_directory = os.environ.get("MAUS_WEB_MEDIA_RAW") if (os.environ.get("MAUS_WEB_MEDIA_RAW") != None) else os.getcwd()
# Default image types of OutputCppRootImage
image_types = ['png']
PolynomialOpticsModel_order = 1
PolynomialOpticsModel_algorithms = ["LeastSquares",
"ConstrainedLeastSquares", "ConstrainedChiSquared",
"SweepingChiSquared", "SweepingChiSquaredWithVariableWalls"]
PolynomialOpticsModel_algorithm = "LeastSquares"
# deltas for numerical derivative calculation of Optics transfer maps
TransferMapOpticsModel_Deltas = {"t":0.01, "E":0.1,
"x":0.1, "Px":0.1,
"y":0.1, "Py":0.01}
root_document_store_timeout = 10
root_document_store_poll_time = 1
geometry_validation = { # see bin/utilities/geometry_validation.py for docs
"file_name":os.path.expandvars("${MAUS_TMP_DIR}/geometry_validation.json"),
"will_plot":True,
"will_track":True,
"z_start":-6000.,
"z_end":6000.,
"x_start":0.,
"x_step":1.,
"y_start":0.,
"y_step":0.,
"n_steps":301,
"plot_formats":["root", "png"],
"1d_material_plot":os.path.expandvars("${MAUS_TMP_DIR}/geometry_validation_materials_1d"),
"2d_material_plot":os.path.expandvars("${MAUS_TMP_DIR}/geometry_validation_materials_2d"),
"1d_volume_plot":os.path.expandvars("${MAUS_TMP_DIR}/geometry_validation_volumes_1d"),
"2d_volume_plot":os.path.expandvars("${MAUS_TMP_DIR}/geometry_validation_volumes_2d"),
"2d_volume_plot_label_size":0.25,
"volume_bounding_box_dump":"geometry_validation_bb_dump.json",
}
# Default location of root file containing PDF histograms used for Global PID
PID_PDFs_file = '%s/src/map/MapCppGlobalPID/PIDhists.root' % os.environ.get("MAUS_ROOT_DIR")
#PID_PDFs_file = '%s/src/map/MapCppGlobalPID/com_pid_hists.root' % os.environ.get("MAUS_ROOT_DIR")
# Tag used by both MapCppGlobalPID and ReduceCppGlobalPID, determines which PDFs to perform PID
# against/which PDFs to produce (in this case, set based upon input MC beam). A typical tag here
# would be the emittance and momentum, e.g. 3-140, 6-200, etc.
pid_beam_setting = "6-200"
# Unique identifier used when creating PDFs in Global PID to distinguish between PDFs for
# the same hypothesis generated at different times. For PDFs to be produced, this must be set.
# Any string can be used but date and time is recommended, by using python datetime module and
# the line unique_identifier = (datetime.datetime.now()).strftime("%Y_%m_%dT%H_%M_%S_%f")
unique_identifier = ""
# Polarity of running mode, used by pid to select whether to run pid against positive or
# negative particles, value can be positive or negative.
pid_beamline_polarity = "positive"
# Bounds set on values of PID variables when running PID
pid_bounds = {
# Bounds for PIDVarA
"minA":20,"maxA":40,
# PIDVarB
"XminB":10, "XmaxB":250, "YminB":20, "YmaxB":40,
# PIDVarC
"XminC":50, "XmaxC":350, "YminC":0, "YmaxC":8000,
# PIDVarD
"minD":0, "maxD":8000,
# PIDVarE
"minE":0, "maxE":1000,
# PIDVarF
"XminF":50, "XmaxF":350, "YminF":0, "YmaxF":1000,
# PIDVarG
"minG":0, "maxG":1,
# PIDVarH
"XminH":50, "XmaxH":350, "YminH":0, "YmaxH":1,
# PIDVarI
"XminI":50, "XmaxI":350, "YminI":0, "YmaxI":140,
# PIDVarJ
"XminJ":50, "XmaxJ":350, "YminJ":0, "YmaxJ":140,
# ComPIDVarA
"minComA":20, "maxComA":50,
# ComPIDVarB
"XminComB":20, "XmaxComB":50, "YminComB":0, "YmaxComB":8000,
# ComPIDVarC
"minComC":0, "maxComC":8000,
# ComPIDVarD
"minComD":0, "maxComD":1000,
# ComPIDVarE
"XminComE":20, "XmaxComE":60, "YminComE":0, "YmaxComE":1000,
# ComPIDVarF
"minComF":0, "maxComF":1,
# ComPIDVarG
"XminComG":20, "XmaxComG":50, "YminComG":0, "YmaxComG":1,
# ComPIDVarH
"XminComH":20, "XmaxComH":50, "YminComH":0, "YmaxComH":140,
# ComPIDVarI
"XminComI":20, "XmaxComI":50, "YminComI":0, "YmaxComI":140
}
# PID MICE configuration, 'step_4' for Step IV running, 'commissioning' for field free commissioning data
pid_config = "step_4"
# PID running mode - selects which PID variables are used. 'online' corresponds to less beam (momentum)
# dependent variables, 'offline' uses all variables and requires that specific PDFs for the beam already
# exist. 'custom' allows user to choose which variables to use, and these should then be set as datacards.
# However it is not recommended to use the custom setting unless you are the person currently developing
# the Global PID.
pid_mode = "offline"
# If pid_mode = "custom", variables to use should be set here as a space separated list, i.e.
# custom_pid_set = "PIDVarA PIDVarC PIDVarD".
custom_pid_set = "PIDVarB"
# PID confidence level- set the margin (in %) between the confidence levels of competing pid hypotheses before they
# are selected as the correct hypothesis
pid_confidence_level = 10
# PID track selection- select which tracks from TrackMatching to perform PID on. Can perform PID on all tracks by
# setting to "all", on through tracks only (constituent tracks will be PID'd, so this excludes orphans) with
# "through" or on all upstream and downstream tracks (ignoring whether tracks have been through-matched) with
# "us_and_ds"
pid_track_selection = "all"
# Determines for which pid hypotheses track matching should be attempted. Default is "all"
# meaning electrons, muons, and pions of both charges (unless tracker recon produces a
# charge hypothesis). Alternatively, force/limit to either one (never several) of
# kEPlus, kEMinus, kMuPlus, kMuMinus, kPiPlus, kPiMinus
track_matching_pid_hypothesis = "all"
# Global track matching tolerances (in mm) for the various subdetectors. KL only provides a
# y coordinate, hence x does not need to be configurable. EMR uses reconstructed error
# so a multiplier is used.
track_matching_tolerances = {
"TOF0t":10.0, # ns between actual and expected TOF0-1 Delta t
"TOF1x":1000.0,
"TOF1y":1000.0,
"TOF2x":1000.0,
"TOF2y":1000.0,
"KLy":1000.0,
"EMRx":1000.0, # Multiplier for the standard tolerance which is the reconstructed error*sqrt(12)
"EMRy":1000.0,
"TOF12maxSpeed":1.0, # fraction of c to calculate travel time between TOFs for through matching
"TOF12minSpeed":0.5,
"TKDpos":1e6, # position
"TKDp":1e6, # momentum
}
# Whether to use energy loss calculations for global track matching
track_matching_energy_loss = "most_probable_energy_loss"
# Whether propagation matching should not be performed if each detector has no more than one hit
track_matching_no_single_event_check = {
"Upstream":False,
"Downstream":False
}
# Controls how geometry lookups (and tracking dynamic stepping) are done:
# "geant4" use full geant4 lookup
# "axial" assume cylindrical symmetry with materials and thicknesses determined by materials detected on-axis
# "geant4_alt" alternative version of the full geant4 lookup
track_matching_geometry_algorithm = "axial"
# Set to false to disable through matching (track matching between TKU and TKD)
# Set to true to enable through matching
track_matching_through_matching = True
# Controls the through matching logic
# no_through_matching - don't do any through matching
# tof12 - Match tracks if they have a tof1 space point, tof2 space point and the
# delta tof12 is compatible with track_matching_tolerances. The
# resultant track is a merge of the upstream and downstream tracks.
# propagate - Match tracks if they have a tku st 1 space point, a tku st2 space
# point and the residual position and momentum between propagated tku
# tracks and tkd tracks is compatible with track_matching_tolerances.
# The resultant track is the upstream track extrapolated to all other
# detectors (so e.g. don't use this track if you want to know what was
# reconstructed in TKD); the downstream track and extrapolation is
# stored in the "constituent tracks" data.
# propagate_requiring_tof12 - if tof12 matching is successful, attempt to do a
# propagate matching. Record the propagate matching if it is successful,
# else the tof12 matching
# propagate_and_tof12 - do a propagate matching and a tof12 matching. Record the
# propagate matching if it is successful, else the tof12 matching
track_matching_through_matching_logic = "propagate_and_tof12"
# Whether residuals should be generated during track matching. In this case,
# track_matching_pid_hypothesis should be set to the most common PID expected in the beam.
# Output files will be placed in the directory in which MAUS is executed.
track_matching_residuals = False
# Choose to add additional z positions for tracking output
# "none": don't add any additional z-positions
# "virtual": use virtual planes as additional z-positions
# Will add track points to the matched tracking output when track propagation crosses
# a virtual plane
track_matching_z_planes = "virtual"
# Whether multiple adjacent cell hits in the KL should be merged into single spacepoints on import
# into the global datastructure
global_merge_kl_cell_hits = True
# Event viewer settings
# EVOnlineSpillStep in an integer value that defines number of spills after which events from one spill will be exported
EVOnlineSpillStep = 100